/linux-6.3-rc2/kernel/sched/ |
A D | topology.c | 398 cpumask_pr_args(cpu_map)); in build_perf_domains() 405 cpumask_pr_args(cpu_map)); in build_perf_domains() 410 for_each_cpu(i, cpu_map) { in build_perf_domains() 1462 __sdt_free(cpu_map); in __free_domain_allocs() 1474 if (__sdt_alloc(cpu_map)) in __visit_domain_allocation_hell() 2321 for_each_cpu(i, cpu_map) { in topology_span_sane() 2360 for_each_cpu(i, cpu_map) { in build_sched_domains() 2383 for_each_cpu(i, cpu_map) { in build_sched_domains() 2400 for_each_cpu(i, cpu_map) { in build_sched_domains() 2467 for_each_cpu(i, cpu_map) { in build_sched_domains() [all …]
|
/linux-6.3-rc2/arch/mips/kernel/ |
A D | cacheinfo.c | 58 static void fill_cpumask_siblings(int cpu, cpumask_t *cpu_map) in fill_cpumask_siblings() argument 64 cpumask_set_cpu(cpu1, cpu_map); in fill_cpumask_siblings() 67 static void fill_cpumask_cluster(int cpu, cpumask_t *cpu_map) in fill_cpumask_cluster() argument 74 cpumask_set_cpu(cpu1, cpu_map); in fill_cpumask_cluster()
|
/linux-6.3-rc2/tools/power/x86/intel-speed-select/ |
A D | isst-config.c | 68 struct _cpu_map *cpu_map; variable 309 if (cpu_map && cpu_map[cpu].initialized) in get_physical_package_id() 310 return cpu_map[cpu].pkg_id; in get_physical_package_id() 333 if (cpu_map && cpu_map[cpu].initialized) in get_physical_core_id() 357 if (cpu_map && cpu_map[cpu].initialized) in get_physical_die_id() 650 cpu_map = calloc(topo_max_cpus, sizeof(*cpu_map)); in create_cpu_map() 651 if (!cpu_map) in create_cpu_map() 701 cpu_map[i].punit_cpu = map.cpu_map[0].physical_cpu; in create_cpu_map() 702 cpu_map[i].punit_cpu_core = (map.cpu_map[0].physical_cpu >> in create_cpu_map() 707 i, cpu_map[i].core_id, cpu_map[i].die_id, in create_cpu_map() [all …]
|
/linux-6.3-rc2/tools/testing/selftests/bpf/progs/ |
A D | freplace_progmap.c | 10 } cpu_map SEC(".maps"); 21 return bpf_redirect_map(&cpu_map, 0, XDP_PASS); in xdp_cpumap_prog()
|
A D | test_xdp_with_cpumap_helpers.c | 13 } cpu_map SEC(".maps"); 18 return bpf_redirect_map(&cpu_map, 1, 0); in xdp_redir_prog()
|
A D | test_xdp_with_cpumap_frags_helpers.c | 13 } cpu_map SEC(".maps");
|
A D | xdp_features.c | 52 } cpu_map SEC(".maps"); 223 return bpf_redirect_map(&cpu_map, 0, 0); in xdp_do_redirect()
|
/linux-6.3-rc2/arch/ia64/mm/ |
A D | discontig.c | 184 unsigned int *cpu_map; in setup_per_cpu_areas() local 194 cpu_map = ai->groups[0].cpu_map; in setup_per_cpu_areas() 208 cpu_map[unit++] = cpu; in setup_per_cpu_areas() 233 cpu = cpu_map[unit]; in setup_per_cpu_areas() 245 gi->cpu_map = &cpu_map[unit]; in setup_per_cpu_areas()
|
A D | contig.c | 115 gi->cpu_map[gi->nr_units++] = cpu; in setup_per_cpu_areas()
|
/linux-6.3-rc2/Documentation/bpf/ |
A D | map_cpumap.rst | 103 ``cpu_map`` and how to redirect packets to a remote CPU using a round robin scheme. 112 } cpu_map SEC(".maps"); 153 return bpf_redirect_map(&cpu_map, cpu_dest, 0); 164 int set_max_cpu_entries(struct bpf_map *cpu_map) 166 if (bpf_map__set_max_entries(cpu_map, libbpf_num_possible_cpus()) < 0) { 167 fprintf(stderr, "Failed to set max entries for cpu_map map: %s",
|
/linux-6.3-rc2/samples/bpf/ |
A D | xdp_redirect_cpu.bpf.c | 15 } cpu_map SEC(".maps"); 164 return bpf_redirect_map(&cpu_map, cpu_dest, 0); in xdp_prognum0_no_touch() 206 return bpf_redirect_map(&cpu_map, cpu_dest, 0); in xdp_prognum1_touch_data() 251 return bpf_redirect_map(&cpu_map, cpu_dest, 0); in xdp_prognum2_round_robin() 317 return bpf_redirect_map(&cpu_map, cpu_dest, 0); in xdp_prognum3_proto_separate() 390 return bpf_redirect_map(&cpu_map, cpu_dest, 0); in xdp_prognum4_ddos_filter_pktgen() 491 return bpf_redirect_map(&cpu_map, cpu_dest, 0); in xdp_prognum5_lb_hash_ip_pairs()
|
A D | xdp_redirect_cpu_user.c | 364 if (bpf_map__set_max_entries(skel->maps.cpu_map, n_cpus) < 0) { in main() 497 ret = bpf_map_get_info_by_fd(bpf_map__fd(skel->maps.cpu_map), &info, &infosz); in main() 506 map_fd = bpf_map__fd(skel->maps.cpu_map); in main()
|
/linux-6.3-rc2/drivers/platform/x86/intel/speed_select_if/ |
A D | isst_if_common.c | 467 struct isst_if_cpu_map *cpu_map; in isst_if_proc_phyid_req() local 469 cpu_map = (struct isst_if_cpu_map *)cmd_ptr; in isst_if_proc_phyid_req() 470 if (cpu_map->logical_cpu >= nr_cpu_ids || in isst_if_proc_phyid_req() 471 cpu_map->logical_cpu >= num_possible_cpus()) in isst_if_proc_phyid_req() 475 cpu_map->physical_cpu = isst_cpu_info[cpu_map->logical_cpu].punit_cpu_id; in isst_if_proc_phyid_req() 598 cmd_cb.offset = offsetof(struct isst_if_cpu_maps, cpu_map); in isst_if_def_ioctl()
|
/linux-6.3-rc2/tools/perf/tests/ |
A D | cpumap.c | 19 struct perf_record_cpu_map *map_event = &event->cpu_map; in process_event_mask() 55 struct perf_record_cpu_map *map_event = &event->cpu_map; in process_event_cpus() 81 struct perf_record_cpu_map *map_event = &event->cpu_map; in process_event_range_cpus()
|
/linux-6.3-rc2/tools/perf/util/ |
A D | mmap.c | 249 const struct perf_cpu_map *cpu_map = NULL; in build_node_mask() local 251 cpu_map = cpu_map__online(); in build_node_mask() 252 if (!cpu_map) in build_node_mask() 255 nr_cpus = perf_cpu_map__nr(cpu_map); in build_node_mask() 257 cpu = perf_cpu_map__cpu(cpu_map, idx); /* map c index to online cpu index */ in build_node_mask()
|
A D | tool.h | 75 cpu_map, member
|
/linux-6.3-rc2/kernel/bpf/ |
A D | cpumap.c | 82 struct bpf_cpu_map_entry __rcu **cpu_map; member 113 cmap->cpu_map = bpf_map_area_alloc(cmap->map.max_entries * in cpu_map_alloc() 116 if (!cmap->cpu_map) { in cpu_map_alloc() 535 old_rcpu = unrcu_pointer(xchg(&cmap->cpu_map[key_cpu], RCU_INITIALIZER(rcpu))); in __cpu_map_entry_replace() 617 rcpu = rcu_dereference_raw(cmap->cpu_map[i]); in cpu_map_free() 624 bpf_map_area_free(cmap->cpu_map); in cpu_map_free() 640 rcpu = rcu_dereference_check(cmap->cpu_map[key], in __cpu_map_lookup_elem()
|
/linux-6.3-rc2/tools/testing/selftests/bpf/prog_tests/ |
A D | xdp_cpumap_attach.c | 35 map_fd = bpf_map__fd(skel->maps.cpu_map); in test_xdp_with_cpumap_helpers() 87 map_fd = bpf_map__fd(skel->maps.cpu_map); in test_xdp_with_cpumap_frags_helpers()
|
/linux-6.3-rc2/mm/ |
A D | percpu.c | 2446 ai->groups[0].cpu_map = ptr; in pcpu_alloc_alloc_info() 2515 if (gi->cpu_map[unit] != NR_CPUS) in pcpu_dump_alloc_info() 2517 cpu_width, gi->cpu_map[unit]); in pcpu_dump_alloc_info() 2662 cpu = gi->cpu_map[i]; in pcpu_setup_first_chunk() 2859 unsigned int *cpu_map; in pcpu_build_alloc_info() local 2951 cpu_map = ai->groups[0].cpu_map; in pcpu_build_alloc_info() 2954 ai->groups[group].cpu_map = cpu_map; in pcpu_build_alloc_info() 2977 gi->cpu_map[gi->nr_units++] = cpu; in pcpu_build_alloc_info() 3089 cpu = gi->cpu_map[i]; in pcpu_embed_first_chunk() 3130 if (gi->cpu_map[i] == NR_CPUS) { in pcpu_embed_first_chunk() [all …]
|
/linux-6.3-rc2/tools/perf/python/ |
A D | twatch.py | 12 cpus = perf.cpu_map()
|
A D | tracepoint.py | 19 cpus = perf.cpu_map()
|
/linux-6.3-rc2/tools/perf/arch/arm/util/ |
A D | cs-etm.c | 799 struct perf_cpu_map *cpu_map; in cs_etm_info_fill() local 814 cpu_map = online_cpus; in cs_etm_info_fill() 825 cpu_map = event_cpus; in cs_etm_info_fill() 828 nr_cpu = perf_cpu_map__nr(cpu_map); in cs_etm_info_fill() 844 if (perf_cpu_map__has(cpu_map, cpu)) in cs_etm_info_fill()
|
/linux-6.3-rc2/include/uapi/linux/ |
A D | isst_if.h | 63 struct isst_if_cpu_map cpu_map[1]; member
|
/linux-6.3-rc2/include/linux/ |
A D | percpu.h | 69 unsigned int *cpu_map; /* unit->cpu map, empty member
|
/linux-6.3-rc2/drivers/hwmon/ |
A D | coretemp.c | 91 u16 cpu_map[NUM_REAL_CORES]; member 521 pdata->cpu_map[index] = topology_core_id(cpu); in create_core_data() 717 if (pd->cpu_map[i] == topology_core_id(cpu)) { in coretemp_cpu_offline()
|