| /tools/lib/perf/ |
| A D | cpumap.c | 20 RC_CHK_ACCESS(map)->nr = nr_cpus; in perf_cpu_map__set_nr() 28 if (nr_cpus == 0) in perf_cpu_map__alloc() 33 cpus->nr = nr_cpus; in perf_cpu_map__alloc() 81 int nr_cpus, nr_cpus_conf; in cpu_map__new_sysconf() local 84 if (nr_cpus < 0) in cpu_map__new_sysconf() 88 if (nr_cpus != nr_cpus_conf) { in cpu_map__new_sysconf() 90 nr_cpus, nr_cpus_conf, nr_cpus); in cpu_map__new_sysconf() 97 for (i = 0; i < nr_cpus; ++i) in cpu_map__new_sysconf() 160 assert(j <= nr_cpus); in cpu_map__trim_new() 170 int i, nr_cpus = 0; in perf_cpu_map__new() local [all …]
|
| /tools/testing/selftests/bpf/prog_tests/ |
| A D | map_lookup_percpu_elem.c | 11 int ret, i, nr_cpus = libbpf_num_possible_cpus(); in test_map_lookup_percpu_elem() local 14 buf = malloc(nr_cpus*sizeof(__u64)); in test_map_lookup_percpu_elem() 18 for (i = 0; i < nr_cpus; i++) in test_map_lookup_percpu_elem() 20 sum = (nr_cpus - 1) * nr_cpus / 2; in test_map_lookup_percpu_elem() 27 skel->rodata->nr_cpus = nr_cpus; in test_map_lookup_percpu_elem()
|
| A D | map_init.c | 10 static int nr_cpus; variable 22 pcpu_map_value_t value[nr_cpus]; in map_populate() 26 for (i = 0; i < nr_cpus; i++) in map_populate() 105 for (i = 0; i < nr_cpus; i++) { in check_values_one_cpu() 131 pcpu_map_value_t value[nr_cpus]; in test_pcpu_map_init() 171 pcpu_map_value_t value[nr_cpus]; in test_pcpu_lru_map_init() 203 nr_cpus = bpf_num_possible_cpus(); in test_map_init() 204 if (nr_cpus <= 1) { in test_map_init()
|
| A D | lookup_and_delete.c | 11 static int nr_cpus; variable 29 __u64 key, value[nr_cpus]; in fill_values_percpu() 32 for (i = 0; i < nr_cpus; i++) in fill_values_percpu() 137 __u64 key, val, value[nr_cpus]; in test_lookup_and_delete_percpu_hash() 156 for (i = 0; i < nr_cpus; i++) { in test_lookup_and_delete_percpu_hash() 223 __u64 key, val, value[nr_cpus]; in test_lookup_and_delete_lru_percpu_hash() 242 for (i = 0; i < nr_cpus; i++) in test_lookup_and_delete_lru_percpu_hash() 252 for (i = 0; i < nr_cpus; i++) { in test_lookup_and_delete_lru_percpu_hash() 281 nr_cpus = bpf_num_possible_cpus(); in test_lookup_and_delete()
|
| A D | perf_buffer.c | 48 int err, on_len, nr_on_cpus = 0, nr_cpus, i, j; in serial_test_perf_buffer() local 56 nr_cpus = libbpf_num_possible_cpus(); in serial_test_perf_buffer() 57 if (CHECK(nr_cpus < 0, "nr_cpus", "err %d\n", nr_cpus)) in serial_test_perf_buffer() 94 for (i = 0; i < nr_cpus; i++) { in serial_test_perf_buffer() 117 for (i = 0, j = 0; i < nr_cpus; i++) { in serial_test_perf_buffer()
|
| A D | xdp_noinline.c | 8 unsigned int nr_cpus = bpf_num_possible_cpus(); in test_xdp_noinline() local 19 } stats[nr_cpus]; in test_xdp_noinline() 66 for (i = 0; i < nr_cpus; i++) { in test_xdp_noinline()
|
| A D | l4lb_all.c | 7 unsigned int nr_cpus = bpf_num_possible_cpus(); in test_l4lb() local 17 } stats[nr_cpus]; in test_l4lb() 79 for (i = 0; i < nr_cpus; i++) { in test_l4lb()
|
| /tools/testing/selftests/kvm/arm64/ |
| A D | vgic_lpi_stress.c | 34 u32 nr_cpus; member 47 .nr_cpus = 1, 68 u32 nr_cpus = test_data.nr_cpus; in guest_setup_its_mappings() local 85 coll_id = (coll_id + 1) % test_data.nr_cpus; in guest_setup_its_mappings() 94 for (i = 0; i < test_data.nr_cpus; i++) in guest_invalidate_all_rdists() 103 gic_init(GIC_V3, test_data.nr_cpus); in guest_setup_gic() 158 sz += (1 + test_data.nr_cpus) * SZ_64K; in setup_memslot() 185 u32 nr_cpus = test_data.nr_cpus; in setup_test_data() local 301 u32 nr_vcpus = test_data.nr_cpus; in run_test() 343 for (i = 0; i < test_data.nr_cpus; i++) in setup_vm() [all …]
|
| /tools/tracing/rtla/src/ |
| A D | timerlat_u.c | 106 static int timerlat_u_send_kill(pid_t *procs, int nr_cpus) in timerlat_u_send_kill() argument 111 for (i = 0; i < nr_cpus; i++) { in timerlat_u_send_kill() 138 int nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in timerlat_u_dispatcher() local 150 procs = calloc(nr_cpus, sizeof(pid_t)); in timerlat_u_dispatcher() 154 for (i = 0; i < nr_cpus; i++) { in timerlat_u_dispatcher() 177 timerlat_u_send_kill(procs, nr_cpus); in timerlat_u_dispatcher() 190 for (i = 0; i < nr_cpus; i++) { in timerlat_u_dispatcher() 204 timerlat_u_send_kill(procs, nr_cpus); in timerlat_u_dispatcher() 212 for (i = 0; i < nr_cpus; i++) { in timerlat_u_dispatcher()
|
| A D | timerlat_hist.c | 48 int nr_cpus; member 94 data->nr_cpus = nr_cpus; in timerlat_alloc_histogram() 205 value_thread[data->nr_cpus], in timerlat_hist_bpf_pull_data() 206 value_user[data->nr_cpus]; in timerlat_hist_bpf_pull_data() 224 data->nr_cpus); in timerlat_hist_bpf_pull_data() 235 data->nr_cpus); in timerlat_hist_bpf_pull_data() 246 data->nr_cpus); in timerlat_hist_bpf_pull_data() 257 data->nr_cpus); in timerlat_hist_bpf_pull_data() 268 data->nr_cpus); in timerlat_hist_bpf_pull_data() 1122 int nr_cpus; in timerlat_init_hist() local [all …]
|
| A D | timerlat_top.c | 46 int nr_cpus; member 71 data->nr_cpus = nr_cpus; in timerlat_alloc_top() 199 data->nr_cpus); in timerlat_top_bpf_pull_data() 210 data->nr_cpus); in timerlat_top_bpf_pull_data() 221 data->nr_cpus); in timerlat_top_bpf_pull_data() 232 data->nr_cpus); in timerlat_top_bpf_pull_data() 243 data->nr_cpus); in timerlat_top_bpf_pull_data() 443 static int nr_cpus = -1; in timerlat_print_stats() local 449 if (nr_cpus == -1) in timerlat_print_stats() 861 int nr_cpus; in timerlat_init_top() local [all …]
|
| A D | osnoise_hist.c | 33 int nr_cpus; member 45 for (cpu = 0; cpu < data->nr_cpus; cpu++) { in osnoise_free_histogram() 72 data->nr_cpus = nr_cpus; in osnoise_alloc_histogram() 78 for (cpu = 0; cpu < nr_cpus; cpu++) { in osnoise_alloc_histogram() 85 for (cpu = 0; cpu < nr_cpus; cpu++) in osnoise_alloc_histogram() 245 for (cpu = 0; cpu < data->nr_cpus; cpu++) { in osnoise_hist_header() 276 for (cpu = 0; cpu < data->nr_cpus; cpu++) { in osnoise_print_summary() 290 for (cpu = 0; cpu < data->nr_cpus; cpu++) { in osnoise_print_summary() 305 for (cpu = 0; cpu < data->nr_cpus; cpu++) { in osnoise_print_summary() 724 int nr_cpus; in osnoise_init_hist() local [all …]
|
| A D | osnoise_top.c | 34 int nr_cpus; member 50 static struct osnoise_top_data *osnoise_alloc_top(int nr_cpus) in osnoise_alloc_top() argument 58 data->nr_cpus = nr_cpus; in osnoise_alloc_top() 61 data->cpu_data = calloc(1, sizeof(*data->cpu_data) * nr_cpus); in osnoise_alloc_top() 229 static int nr_cpus = -1; in osnoise_print_stats() local 232 if (nr_cpus == -1) in osnoise_print_stats() 233 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in osnoise_print_stats() 240 for (i = 0; i < nr_cpus; i++) { in osnoise_print_stats() 552 int nr_cpus; in osnoise_init_top() local 554 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in osnoise_init_top() [all …]
|
| A D | utils.c | 106 int nr_cpus; in parse_cpu_set() local 112 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in parse_cpu_set() 116 if (cpu < 0 || (!cpu && *p != '0') || cpu >= nr_cpus) in parse_cpu_set() 124 if (end_cpu < cpu || (!end_cpu && *p != '0') || end_cpu >= nr_cpus) in parse_cpu_set() 540 int nr_cpus; in save_cpu_idle_disable_state() local 548 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in save_cpu_idle_disable_state() 549 saved_cpu_idle_disable_state = calloc(nr_cpus, sizeof(unsigned int *)); in save_cpu_idle_disable_state() 625 int nr_cpus; in free_cpu_idle_disable_states() local 630 nr_cpus = sysconf(_SC_NPROCESSORS_CONF); in free_cpu_idle_disable_states() 632 for (cpu = 0; cpu < nr_cpus; cpu++) { in free_cpu_idle_disable_states()
|
| /tools/testing/selftests/bpf/map_tests/ |
| A D | array_map_batch_ops.c | 13 static int nr_cpus; variable 28 cpu_offset = i * nr_cpus; in map_batch_update() 29 for (j = 0; j < nr_cpus; j++) in map_batch_update() 49 cpu_offset = i * nr_cpus; in map_batch_verify() 50 for (j = 0; j < nr_cpus; j++) { in map_batch_verify() 90 value_size *= nr_cpus; in __test_map_lookup_and_update_batch() 158 nr_cpus = libbpf_num_possible_cpus(); in test_array_map_batch_ops() 160 CHECK(nr_cpus < 0, "nr_cpus checking", in test_array_map_batch_ops()
|
| /tools/testing/selftests/bpf/progs/ |
| A D | test_map_lookup_percpu_elem.c | 10 const volatile int nr_cpus; variable 60 bpf_loop(nr_cpus, read_percpu_elem_callback, &map_ctx, 0); in sysenter_getuid() 65 bpf_loop(nr_cpus, read_percpu_elem_callback, &map_ctx, 0); in sysenter_getuid() 70 bpf_loop(nr_cpus, read_percpu_elem_callback, &map_ctx, 0); in sysenter_getuid()
|
| /tools/testing/selftests/kvm/lib/arm64/ |
| A D | gic.c | 25 static void gic_dist_init(enum gic_type type, unsigned int nr_cpus) in gic_dist_init() argument 42 gic_ops->gic_init(nr_cpus); in gic_dist_init() 51 void gic_init(enum gic_type type, unsigned int nr_cpus) in gic_init() argument 56 GUEST_ASSERT(nr_cpus); in gic_init() 58 gic_dist_init(type, nr_cpus); in gic_init()
|
| /tools/testing/selftests/bpf/ |
| A D | test_lru_map.c | 26 static int nr_cpus; variable 96 unsigned long long value0[nr_cpus], value1[nr_cpus]; in map_subset() 127 while (next < nr_cpus) { in sched_next_online() 144 return (map_size / nr_cpus) / 2; in __tgt_size() 150 return tgt_free * nr_cpus * 2; in __map_size() 163 unsigned long long key, value[nr_cpus]; in test_lru_sanity0() 322 unsigned long long key, value[nr_cpus]; in test_lru_sanity2() 501 3 * tgt_free * nr_cpus); in test_lru_sanity4() 860 nr_cpus = bpf_num_possible_cpus(); in main() 861 assert(nr_cpus != -1); in main() [all …]
|
| /tools/testing/selftests/sched_ext/ |
| A D | dsp_local_on.bpf.c | 9 const volatile s32 nr_cpus; variable 46 if (p->nr_cpus_allowed == nr_cpus && !is_migration_disabled(p)) in BPF_STRUCT_OPS() 47 target = bpf_get_prandom_u32() % nr_cpus; in BPF_STRUCT_OPS()
|
| /tools/perf/util/ |
| A D | top.c | 31 int nr_cpus; in perf_top__header_snprintf() local 97 nr_cpus = perf_cpu_map__nr(top->evlist->core.user_requested_cpus); in perf_top__header_snprintf() 100 nr_cpus > 1 ? "s" : "", in perf_top__header_snprintf() 107 nr_cpus, nr_cpus > 1 ? "s" : ""); in perf_top__header_snprintf()
|
| A D | bpf_kwork_top.c | 126 int idx, nr_cpus, fd; in setup_filters() local 142 nr_cpus = libbpf_num_possible_cpus(); in setup_filters() 146 if (cpu.cpu >= nr_cpus) { in setup_filters() 270 int i, fd, nr_cpus; in perf_kwork__top_read_bpf() local 280 nr_cpus = libbpf_num_possible_cpus(); in perf_kwork__top_read_bpf() 281 data = calloc(nr_cpus, sizeof(struct work_data)); in perf_kwork__top_read_bpf() 292 for (i = 0; i < nr_cpus; i++) { in perf_kwork__top_read_bpf()
|
| A D | svghelper.c | 701 int *pos, int nr_cpus) in scan_thread_topology() argument 724 scan_thread_topology(map, t, cpu, &pos, nr_cpus); in scan_core_topology() 727 static int str_to_bitmap(char *s, cpumask_t *b, int nr_cpus) in str_to_bitmap() argument 738 if (cpu.cpu >= nr_cpus) { in str_to_bitmap() 753 int i, nr_cpus; in svg_build_topology_map() local 758 nr_cpus = min(env->nr_cpus_online, MAX_NR_CPUS); in svg_build_topology_map() 774 if (str_to_bitmap(sib_core, &t.sib_core[i], nr_cpus)) { in svg_build_topology_map() 783 if (str_to_bitmap(sib_thr, &t.sib_thr[i], nr_cpus)) { in svg_build_topology_map() 791 topology_map = malloc(sizeof(int) * nr_cpus); in svg_build_topology_map() 797 for (i = 0; i < nr_cpus; i++) in svg_build_topology_map() [all …]
|
| /tools/lib/perf/include/internal/ |
| A D | cpumap.h | 24 struct perf_cpu_map *perf_cpu_map__alloc(int nr_cpus); 28 void perf_cpu_map__set_nr(struct perf_cpu_map *map, int nr_cpus);
|
| /tools/sched_ext/ |
| A D | scx_simple.c | 43 int nr_cpus = libbpf_num_possible_cpus(); in read_stats() local 44 __u64 cnts[2][nr_cpus]; in read_stats() 56 for (cpu = 0; cpu < nr_cpus; cpu++) in read_stats()
|
| /tools/testing/selftests/rcutorture/bin/ |
| A D | functions.sh | 63 if test "$3" -gt "$nr_cpus" 65 echo $nr_cpus
|