| /tools/testing/selftests/rseq/ |
| A D | basic_test.c | 18 cpu_set_t affinity, test_affinity; in test_cpu_pointer() local 21 sched_getaffinity(0, sizeof(affinity), &affinity); in test_cpu_pointer() 24 if (CPU_ISSET(i, &affinity)) { in test_cpu_pointer() 39 sched_setaffinity(0, sizeof(affinity), &affinity); in test_cpu_pointer()
|
| /tools/perf/util/ |
| A D | affinity.h | 8 struct affinity { struct 14 void affinity__cleanup(struct affinity *a); argument 15 void affinity__set(struct affinity *a, int cpu); 16 int affinity__setup(struct affinity *a);
|
| A D | affinity.c | 25 int affinity__setup(struct affinity *a) in affinity__setup() 49 void affinity__set(struct affinity *a, int cpu) in affinity__set() 72 static void __affinity__cleanup(struct affinity *a) in __affinity__cleanup() 82 void affinity__cleanup(struct affinity *a) in affinity__cleanup()
|
| A D | mmap.c | 97 static int perf_mmap__aio_bind(struct mmap *map, int idx, struct perf_cpu cpu, int affinity) in perf_mmap__aio_bind() argument 105 if (affinity != PERF_AFFINITY_SYS && cpu__max_node() > 1) { in perf_mmap__aio_bind() 141 struct perf_cpu cpu __maybe_unused, int affinity __maybe_unused) in perf_mmap__aio_bind() 175 ret = perf_mmap__aio_bind(map, i, map->core.cpu, mp->affinity); in perf_mmap__aio_mmap() 268 if (mp->affinity == PERF_AFFINITY_NODE && cpu__max_node() > 1) in perf_mmap__setup_affinity_mask() 270 else if (mp->affinity == PERF_AFFINITY_CPU) in perf_mmap__setup_affinity_mask() 284 if (mp->affinity != PERF_AFFINITY_SYS && in mmap__mmap()
|
| A D | evlist.h | 196 int affinity, int flush, int comp_level); 363 struct affinity *affinity; member 376 #define evlist__for_each_cpu(evlist_cpu_itr, evlist, affinity) \ argument 377 for ((evlist_cpu_itr) = evlist__cpu_begin(evlist, affinity); \ 382 struct evlist_cpu_iterator evlist__cpu_begin(struct evlist *evlist, struct affinity *affinity);
|
| A D | evlist.c | 362 .affinity = affinity, in evlist__cpu_begin() 370 if (itr.affinity) { in evlist__cpu_begin() 401 if (evlist_cpu_itr->affinity) in evlist_cpu_iterator__next() 447 struct affinity saved_affinity, *affinity = NULL; in __evlist__disable() local 454 affinity = &saved_affinity; in __evlist__disable() 477 affinity__cleanup(affinity); in __evlist__disable() 517 struct affinity saved_affinity, *affinity = NULL; in __evlist__enable() local 523 affinity = &saved_affinity; in __evlist__enable() 536 affinity__cleanup(affinity); in __evlist__enable() 964 .affinity = affinity, in evlist__mmap_ex() [all …]
|
| A D | mmap.h | 48 int nr_cblocks, affinity, flush, comp_level; member
|
| A D | record.h | 73 int affinity; member
|
| A D | Build | 102 perf-util-y += affinity.o
|
| /tools/virtio/ringtest/ |
| A D | run-on-all.sh | 20 "$@" --host-affinity $HOST_AFFINITY --guest-affinity $cpu 24 "$@" --host-affinity $HOST_AFFINITY
|
| /tools/testing/selftests/rcutorture/bin/ |
| A D | kvm-test-1-run-batch.sh | 65 print "echo No CPU-affinity information, so no taskset command."; 71 print "echo " scenario ": Bogus CPU-affinity information, so no taskset command.";
|
| A D | kvm.sh | 225 --no-affinity)
|
| /tools/perf/ |
| A D | builtin-record.c | 100 struct mmap_cpu_mask affinity; member 1334 opts->nr_cblocks, opts->affinity, in record__mmap_evlist() 1543 thread->mask->affinity.nbits)) { in record__adjust_affinity() 1544 bitmap_zero(thread->mask->affinity.bits, thread->mask->affinity.nbits); in record__adjust_affinity() 1545 bitmap_or(thread->mask->affinity.bits, thread->mask->affinity.bits, in record__adjust_affinity() 3058 opts->affinity = PERF_AFFINITY_NODE; in record__parse_affinity() 3060 opts->affinity = PERF_AFFINITY_CPU; in record__parse_affinity() 3087 mask->affinity.bits = NULL; in record__thread_mask_alloc() 3807 if (!bitmap_and(thread_mask.affinity.bits, thread_mask.affinity.bits, in record__init_thread_masks_spec() 3821 if (bitmap_intersects(thread_mask.affinity.bits, full_mask.affinity.bits, in record__init_thread_masks_spec() [all …]
|
| A D | builtin-stat.c | 351 struct affinity saved_affinity, *affinity; in read_affinity_counters() local 357 affinity = NULL; in read_affinity_counters() 361 affinity = &saved_affinity; in read_affinity_counters() 363 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in read_affinity_counters() 372 if (affinity) in read_affinity_counters() 698 struct affinity saved_affinity, *affinity = NULL; in __run_perf_stat() local 715 affinity = &saved_affinity; in __run_perf_stat() 730 evlist__for_each_cpu(evlist_cpu_itr, evsel_list, affinity) { in __run_perf_stat() 821 affinity__cleanup(affinity); in __run_perf_stat() 822 affinity = NULL; in __run_perf_stat() [all …]
|
| /tools/testing/selftests/bpf/ |
| A D | bench.c | 19 .affinity = false, 346 env.affinity = true; in parse_arg() 352 env.affinity = true; in parse_arg() 360 env.affinity = true; in parse_arg() 681 if (env.affinity) in setup_benchmark() 704 if (env.affinity) in setup_benchmark()
|
| A D | bench.h | 27 bool affinity; member
|
| /tools/testing/selftests/seccomp/ |
| A D | seccomp_benchmark.c | 141 void affinity(void) in affinity() function 207 affinity(); in main()
|
| /tools/testing/selftests/ublk/ |
| A D | kublk.c | 323 cpu_set_t *affinity; in ublk_ctrl_dump() local 332 ret = ublk_ctrl_get_affinity(dev, &affinity); in ublk_ctrl_dump() 345 if (affinity) { in ublk_ctrl_dump() 350 ublk_print_cpu_set(&affinity[i], buf, sizeof(buf)); in ublk_ctrl_dump() 354 free(affinity); in ublk_ctrl_dump() 834 cpu_set_t *affinity; member 854 if (info->affinity) in ublk_io_handler_fn() 855 ublk_thread_set_sched_affinity(t, info->affinity); in ublk_io_handler_fn() 964 tinfo[i].affinity = &affinity_buf[i]; in ublk_start_daemon()
|
| /tools/testing/selftests/bpf/benchs/ |
| A D | bench_bpf_hashmap_lookup.c | 255 if (env.affinity) in hashmap_report_final()
|
| A D | bench_sockmap.c | 238 !env.affinity) in validate()
|
| /tools/perf/Documentation/ |
| A D | perf-record.txt | 575 --affinity=mode:: 576 Set affinity mask of trace reading thread according to the policy defined by 'mode' value: 578 - node - thread affinity mask is set to NUMA node cpu mask of the processed mmap buffer 579 - cpu - thread affinity mask is set to cpu of the processed mmap buffer 797 define CPUs to be monitored by a thread and affinity mask of that thread 800 <cpus mask 1>/<affinity mask 1>:<cpus mask 2>/<affinity mask 2>:... 802 CPUs or affinity masks must not overlap with other corresponding masks. 811 the first thread monitors CPUs 0 and 2-4 with the affinity mask 2-4, 812 the second monitors CPUs 1 and 5-7 with the affinity mask 5-7.
|