Home
last modified time | relevance | path

Searched refs:nthreads (Results 1 – 25 of 34) sorted by relevance

12

/tools/testing/selftests/mm/
A Dmigration.c30 int nthreads; in FIXTURE() local
40 self->nthreads = numa_num_task_cpus() - 1; in FIXTURE_SETUP()
137 for (i = 0; i < self->nthreads - 1; i++)
142 for (i = 0; i < self->nthreads - 1; i++)
177 for (i = 0; i < self->nthreads - 1; i++)
202 for (i = 0; i < self->nthreads - 1; i++)
207 for (i = 0; i < self->nthreads - 1; i++)
249 for (i = 0; i < self->nthreads - 1; i++)
269 for (i = 0; i < self->nthreads - 1; i++)
274 for (i = 0; i < self->nthreads - 1; i++)
[all …]
A Dgup_test.c95 int filed, i, opt, nr_pages = 1, thp = -1, write = 1, nthreads = 1, ret; in main() local
132 nthreads = atoi(optarg); in main()
205 ksft_set_plan(nthreads); in main()
258 tid = malloc(sizeof(pthread_t) * nthreads); in main()
260 for (i = 0; i < nthreads; i++) { in main()
264 for (i = 0; i < nthreads; i++) { in main()
A Dpagemap_ioctl.c1410 wp_init(mem, 0x1000 * nthreads * pages_per_thread); in transact_test()
1411 wp_addr_range(mem, 0x1000 * nthreads * pages_per_thread); in transact_test()
1413 memset(mem, 0, 0x1000 * nthreads * pages_per_thread); in transact_test()
1421 for (i = 0; i < nthreads; ++i) in transact_test()
1439 if (count > nthreads * access_per_thread) in transact_test()
1441 count, nthreads * access_per_thread, i); in transact_test()
1446 if (c > nthreads * access_per_thread) { in transact_test()
1451 if (count != nthreads * access_per_thread) { in transact_test()
1464 if (count < nthreads * access_per_thread) { in transact_test()
1466 nthreads * access_per_thread); in transact_test()
[all …]
/tools/perf/bench/
A Dfutex-requeue.c79 params.nthreads, in print_summary()
134 threads_starting = params.nthreads; in block_threads()
141 for (i = 0; i < params.nthreads; i++) { in block_threads()
194 if (!params.nthreads) in bench_futex_requeue()
204 if (params.nrequeue > params.nthreads) in bench_futex_requeue()
205 params.nrequeue = params.nthreads; in bench_futex_requeue()
208 params.nrequeue = params.nthreads; in bench_futex_requeue()
278 params.nthreads, in bench_futex_requeue()
285 params.nthreads, in bench_futex_requeue()
294 if (params.nthreads != nrequeued) in bench_futex_requeue()
[all …]
A Dfutex-wake.c55 OPT_UINTEGER('t', "threads", &params.nthreads, "Specify amount of threads"),
95 params.nthreads, in print_summary()
107 threads_starting = params.nthreads; in block_threads()
114 for (i = 0; i < params.nthreads; i++) { in block_threads()
169 if (!params.nthreads) in bench_futex_wake()
170 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_wake()
172 worker = calloc(params.nthreads, sizeof(*worker)); in bench_futex_wake()
181 getpid(), params.nthreads, params.fshared ? "shared":"private", in bench_futex_wake()
208 while (nwoken != params.nthreads) in bench_futex_wake()
219 j + 1, nwoken, params.nthreads, in bench_futex_wake()
[all …]
A Dbreakpoint.c22 unsigned int nthreads; member
26 .nthreads = 1,
97 for (i = 0; i < thread_params.nthreads; i++) { in breakpoint_thread()
102 futex_wake(&done, thread_params.nthreads, 0); in breakpoint_thread()
103 for (i = 0; i < thread_params.nthreads; i++) in breakpoint_thread()
164 thread_params.nthreads * thread_params.nparallel); in bench_breakpoint_thread()
200 unsigned int i, nthreads, result_usec, done = 0; in bench_breakpoint_enable() local
219 nthreads = enable_params.npassive + enable_params.nactive; in bench_breakpoint_enable()
220 threads = calloc(nthreads, sizeof(threads[0])); in bench_breakpoint_enable()
224 for (i = 0; i < nthreads; i++) { in bench_breakpoint_enable()
[all …]
A Dfutex-wake-parallel.c158 threads_starting = params.nthreads; in block_threads()
165 for (i = 0; i < params.nthreads; i++) { in block_threads()
206 params.nthreads, waketime_avg / USEC_PER_MSEC, in print_run()
221 params.nthreads, in print_summary()
275 if (!params.nthreads) in bench_futex_wake_parallel()
276 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_wake_parallel()
279 if (params.nwakes > params.nthreads || in bench_futex_wake_parallel()
281 params.nwakes = params.nthreads; in bench_futex_wake_parallel()
283 if (params.nthreads % params.nwakes) in bench_futex_wake_parallel()
289 nwakes = params.nthreads/params.nwakes; in bench_futex_wake_parallel()
[all …]
A Dfutex-lock-pi.c50 OPT_UINTEGER('t', "threads", &params.nthreads, "Specify amount of threads"),
131 threads_starting = params.nthreads; in create_threads()
137 for (i = 0; i < params.nthreads; i++) { in create_threads()
192 if (!params.nthreads) in bench_futex_lock_pi()
193 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_lock_pi()
195 worker = calloc(params.nthreads, sizeof(*worker)); in bench_futex_lock_pi()
203 getpid(), params.nthreads, params.runtime); in bench_futex_lock_pi()
211 threads_starting = params.nthreads; in bench_futex_lock_pi()
225 for (i = 0; i < params.nthreads; i++) { in bench_futex_lock_pi()
236 for (i = 0; i < params.nthreads; i++) { in bench_futex_lock_pi()
A Dfutex-hash.c59 OPT_UINTEGER('t', "threads", &params.nthreads, "Specify amount of threads"),
159 if (!params.nthreads) /* default to the number of CPUs */ in bench_futex_hash()
160 params.nthreads = perf_cpu_map__nr(cpu); in bench_futex_hash()
162 worker = calloc(params.nthreads, sizeof(*worker)); in bench_futex_hash()
171 … getpid(), params.nthreads, params.nfutexes, params.fshared ? "shared":"private", params.runtime); in bench_futex_hash()
178 threads_starting = params.nthreads; in bench_futex_hash()
187 for (i = 0; i < params.nthreads; i++) { in bench_futex_hash()
221 for (i = 0; i < params.nthreads; i++) { in bench_futex_hash()
232 for (i = 0; i < params.nthreads; i++) { in bench_futex_hash()
A Depoll-ctl.c36 static unsigned int nthreads = 0; variable
240 for (i = 0; i < nthreads; i++) { in do_threads()
349 if (!nthreads) in bench_epoll_ctl()
350 nthreads = perf_cpu_map__nr(cpu); in bench_epoll_ctl()
352 worker = calloc(nthreads, sizeof(*worker)); in bench_epoll_ctl()
358 rl.rlim_cur = rl.rlim_max = nfds * nthreads * 2 + 50; in bench_epoll_ctl()
366 getpid(), nthreads, nfds, nsecs); in bench_epoll_ctl()
375 threads_starting = nthreads; in bench_epoll_ctl()
391 for (i = 0; i < nthreads; i++) { in bench_epoll_ctl()
402 for (i = 0; i < nthreads; i++) { in bench_epoll_ctl()
[all …]
A Depoll-wait.c91 static unsigned int nthreads = 0; variable
317 for (i = 0; i < nthreads; i++) { in do_threads()
394 shuffle((void *)worker, nthreads, sizeof(*worker)); in writerfn()
397 for (i = 0; i < nthreads; i++) { in writerfn()
473 if (!nthreads) in bench_epoll_wait()
474 nthreads = perf_cpu_map__nr(cpu) - 1; in bench_epoll_wait()
476 worker = calloc(nthreads, sizeof(*worker)); in bench_epoll_wait()
483 rl.rlim_cur = rl.rlim_max = nfds * nthreads * 2 + 50; in bench_epoll_wait()
498 threads_starting = nthreads; in bench_epoll_wait()
539 for (i = 0; i < nthreads; i++) { in bench_epoll_wait()
[all …]
/tools/testing/selftests/bpf/prog_tests/
A Darena_spin_lock.c54 int nthreads; in test_arena_spin_lock_size() local
57 nthreads = MIN(get_nprocs(), ARRAY_SIZE(thread_id)); in test_arena_spin_lock_size()
58 if (nthreads < 2) { in test_arena_spin_lock_size()
71 skel->bss->limit = repeat * nthreads; in test_arena_spin_lock_size()
73 ASSERT_OK(pthread_barrier_init(&barrier, NULL, nthreads), "barrier init"); in test_arena_spin_lock_size()
76 for (i = 0; i < nthreads; i++) { in test_arena_spin_lock_size()
82 for (i = 0; i < nthreads; i++) { in test_arena_spin_lock_size()
89 ASSERT_EQ(skel->bss->counter, repeat * nthreads, "check counter value"); in test_arena_spin_lock_size()
/tools/perf/util/
A Dcounts.c10 struct perf_counts *perf_counts__new(int ncpus, int nthreads) in perf_counts__new() argument
17 values = xyarray__new(ncpus, nthreads, sizeof(struct perf_counts_values)); in perf_counts__new()
25 values = xyarray__new(ncpus, nthreads, sizeof(bool)); in perf_counts__new()
61 int nthreads = perf_thread_map__nr(evsel->core.threads); in evsel__alloc_counts() local
63 evsel->counts = perf_counts__new(perf_cpu_map__nr(cpus), nthreads); in evsel__alloc_counts()
A Dtool_pmu.c203 int nthreads) in evsel__tool_pmu_prepare_open() argument
209 nthreads, in evsel__tool_pmu_prepare_open()
224 int pid = -1, idx = 0, thread = 0, nthreads, err = 0, old_errno; in evsel__tool_pmu_open() local
239 nthreads = perf_thread_map__nr(threads); in evsel__tool_pmu_open()
241 for (thread = 0; thread < nthreads; thread++) { in evsel__tool_pmu_open()
242 if (thread >= nthreads) in evsel__tool_pmu_open()
301 thread = nthreads; in evsel__tool_pmu_open()
A Dcounts.h37 struct perf_counts *perf_counts__new(int ncpus, int nthreads);
A Dtool_pmu.h48 int nthreads);
/tools/testing/selftests/ublk/
A Dtest_stress_04.sh43 ublk_io_and_kill_daemon 8G -t null -q 4 --nthreads 8 --per_io_tasks &
44 ublk_io_and_kill_daemon 256M -t loop -q 4 --nthreads 8 --per_io_tasks "${UBLK_BACKFILES[0]}" &
45 …ublk_io_and_kill_daemon 256M -t stripe -q 4 --nthreads 8 --per_io_tasks "${UBLK_BACKFILES[1]}" "${…
46 ublk_io_and_kill_daemon 8G -t null -q 4 --nthreads 8 --per_io_tasks &
A Dtest_stress_03.sh46 ublk_io_and_remove 8G -t null -q 4 --auto_zc --nthreads 8 --per_io_tasks &
47 ublk_io_and_remove 256M -t loop -q 4 --auto_zc --nthreads 8 --per_io_tasks "${UBLK_BACKFILES[0]}" &
48 …ublk_io_and_remove 256M -t stripe -q 4 --auto_zc --nthreads 8 --per_io_tasks "${UBLK_BACKFILES[1]}…
49 ublk_io_and_remove 8G -t null -q 4 -z --auto_zc --auto_zc_fallback --nthreads 8 --per_io_tasks &
A Dtest_stress_05.sh73 ublk_io_and_remove 8G -t null -q 4 --nthreads 8 --per_io_tasks -r 1 -i "$reissue" &
74 …ublk_io_and_remove 256M -t loop -q 4 --nthreads 8 --per_io_tasks -r 1 -i "$reissue" "${UBLK_BACKFI…
75 ublk_io_and_remove 8G -t null -q 4 --nthreads 8 --per_io_tasks -r 1 -i "$reissue" &
A Dkublk.c950 for (i = 0; i < dev->nthreads; i++) { in ublk_start_daemon()
963 if (dev->nthreads == dinfo->nr_hw_queues) in ublk_start_daemon()
970 for (i = 0; i < dev->nthreads; i++) in ublk_start_daemon()
994 for (i = 0; i < dev->nthreads; i++) in ublk_start_daemon()
1104 unsigned nthreads = ctx->nthreads; in __cmd_dev_add() local
1131 if (!nthreads) in __cmd_dev_add()
1132 nthreads = nr_queues; in __cmd_dev_add()
1134 if (nthreads > UBLK_MAX_THREADS) { in __cmd_dev_add()
1136 __func__, nthreads, UBLK_MAX_THREADS); in __cmd_dev_add()
1144 __func__, nthreads, nr_queues); in __cmd_dev_add()
[all …]
A Dtest_generic_12.sh16 dev_id=$(_add_ublk_dev -t null -q 4 -d 16 --nthreads $NTHREADS --per_io_tasks)
/tools/testing/selftests/mm/page_frag/
A Dpage_frag_test.c20 static atomic_t nthreads; variable
68 if (atomic_dec_and_test(&nthreads)) in page_frag_pop_thread()
115 if (atomic_dec_and_test(&nthreads)) in page_frag_push_thread()
130 atomic_set(&nthreads, 2); in page_frag_test_init()
/tools/lib/perf/include/internal/
A Devsel.h137 int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads);
143 int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads);
/tools/lib/perf/
A Devsel.c65 int perf_evsel__alloc_fd(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_fd() argument
67 evsel->fd = xyarray__new(ncpus, nthreads, sizeof(int)); in perf_evsel__alloc_fd()
73 for (thread = 0; thread < nthreads; thread++) { in perf_evsel__alloc_fd()
85 static int perf_evsel__alloc_mmap(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_mmap() argument
87 evsel->mmap = xyarray__new(ncpus, nthreads, sizeof(struct perf_mmap)); in perf_evsel__alloc_mmap()
526 int perf_evsel__alloc_id(struct perf_evsel *evsel, int ncpus, int nthreads) in perf_evsel__alloc_id() argument
528 if (ncpus == 0 || nthreads == 0) in perf_evsel__alloc_id()
531 evsel->sample_id = xyarray__new(ncpus, nthreads, sizeof(struct perf_sample_id)); in perf_evsel__alloc_id()
535 evsel->id = zalloc(ncpus * nthreads * sizeof(u64)); in perf_evsel__alloc_id()
/tools/testing/selftests/filesystems/binderfs/
A Dbinderfs_test.c403 int i, j, k, nthreads; in TEST() local
453 nthreads = get_nprocs_conf(); in TEST()
454 if (nthreads > DEFAULT_THREADS) in TEST()
455 nthreads = DEFAULT_THREADS; in TEST()
460 for (i = 0; i < nthreads; i++) { in TEST()

Completed in 35 milliseconds

12