| /linux/drivers/cpuidle/governors/ |
| A D | teo.c | 150 if (cpu_data->time_span_ns >= cpu_data->sleep_length_ns) { in teo_update() 180 cpu_data->total = 0; in teo_update() 211 cpu_data->tick_hits -= cpu_data->tick_hits >> DECAY_SHIFT; in teo_update() 213 cpu_data->total += cpu_data->tick_hits; in teo_update() 218 cpu_data->tick_hits += PULSE; in teo_update() 236 cpu_data->total += PULSE; in teo_update() 304 cpu_data->sleep_length_ns = KTIME_MAX; in teo_select() 472 tick_intercept_sum > cpu_data->total / 2 + cpu_data->total / 8) in teo_select() 517 cpu_data->time_span_ns = cpu_data->sleep_length_ns; in teo_reflect() 519 cpu_data->time_span_ns = local_clock() - cpu_data->time_span_ns; in teo_reflect() [all …]
|
| /linux/drivers/cpufreq/ |
| A D | cppc_cpufreq.c | 75 struct cppc_cpudata *cpu_data; member 107 struct cppc_cpudata *cpu_data; in cppc_scale_freq_workfn() local 112 cpu_data = cppc_fi->cpu_data; in cppc_scale_freq_workfn() 387 struct cppc_cpudata *cpu_data; in get_perf_level_count() local 564 if (!cpu_data) in cppc_cpufreq_get_cpu_data() 584 return cpu_data; in cppc_cpufreq_get_cpu_data() 589 kfree(cpu_data); in cppc_cpufreq_get_cpu_data() 598 list_del(&cpu_data->node); in cppc_cpufreq_put_cpu_data() 600 kfree(cpu_data); in cppc_cpufreq_put_cpu_data() 612 if (!cpu_data) { in cppc_cpufreq_cpu_init() [all …]
|
| A D | tegra194-cpufreq.c | 79 struct tegra_cpu_data *cpu_data; member 128 *ndiv = readl(data->cpu_data[cpu].freq_core_reg) & NDIV_MASK; in tegra234_get_cpu_ndiv() 139 writel(ndiv, data->cpu_data[cpu].freq_core_reg); in tegra234_set_cpu_ndiv() 158 data->cpu_data[c->cpu].cpuid); in tegra234_read_counters() 388 u32 clusterid = data->cpu_data[cpu].clusterid; in tegra194_get_speed() 504 u32 clusterid = data->cpu_data[policy->cpu].clusterid; in tegra194_cpufreq_init() 701 data->cpu_data[cpu].cpuid = cpuid; in tegra194_cpufreq_store_physids() 702 data->cpu_data[cpu].clusterid = clusterid; in tegra194_cpufreq_store_physids() 742 data->cpu_data = devm_kcalloc(&pdev->dev, data->soc->num_clusters * in tegra194_cpufreq_probe() 744 sizeof(*data->cpu_data), GFP_KERNEL); in tegra194_cpufreq_probe() [all …]
|
| /linux/arch/mips/include/asm/ |
| A D | cpu-features.h | 17 #define __ase(ase) (cpu_data[0].ases & (ase)) 18 #define __isa(isa) (cpu_data[0].isa_level & (isa)) 19 #define __opt(opt) (cpu_data[0].options & (opt)) 486 # define cpu_vmbits cpu_data[0].vmbits 508 #define cpu_dcache_line_size() cpu_data[0].dcache.linesz 511 #define cpu_icache_line_size() cpu_data[0].icache.linesz 514 #define cpu_scache_line_size() cpu_data[0].scache.linesz 517 #define cpu_tcache_line_size() cpu_data[0].tcache.linesz 666 #define cpu_guest_has_conf1 (cpu_data[0].guest.conf & (1 << 1)) 669 #define cpu_guest_has_conf2 (cpu_data[0].guest.conf & (1 << 2)) [all …]
|
| A D | cpu-info.h | 119 extern struct cpuinfo_mips cpu_data[]; 120 #define current_cpu_data cpu_data[smp_processor_id()] 121 #define raw_current_cpu_data cpu_data[raw_smp_processor_id()] 122 #define boot_cpu_data cpu_data[0] 183 struct cpuinfo_mips *infoa = &cpu_data[cpua]; in cpus_are_siblings() 184 struct cpuinfo_mips *infob = &cpu_data[cpub]; in cpus_are_siblings()
|
| A D | mmu_context.h | 96 unsigned long asid_mask = cpu_asid_mask(&cpu_data[cpu]); in asid_version_mask() 123 #define asid_cache(cpu) (cpu_data[cpu].asid_cache) 125 (cpu_context((cpu), (mm)) & cpu_asid_mask(&cpu_data[cpu])) 211 write_c0_memorymapid(ctx & cpu_asid_mask(&cpu_data[cpu])); in drop_mmu_context()
|
| /linux/tools/tracing/rtla/src/ |
| A D | timerlat_top.c | 88 free(data->cpu_data); in timerlat_free_top() 107 data->cpu_data = calloc(1, sizeof(*data->cpu_data) * nr_cpus); in timerlat_alloc_top() 108 if (!data->cpu_data) in timerlat_alloc_top() 138 struct timerlat_top_cpu *cpu_data = &data->cpu_data[cpu]; in timerlat_top_update_sum() local 165 struct timerlat_top_cpu *cpu_data = &data->cpu_data[cpu]; in timerlat_top_update() local 168 cpu_data->irq_count++; in timerlat_top_update() 180 cpu_data->user_count++; in timerlat_top_update() 267 struct timerlat_top_cpu *cpu_data = &data->cpu_data[cpu]; in timerlat_top_print() local 277 if (!cpu_data->irq_count && !cpu_data->thread_count) in timerlat_top_print() 300 (cpu_data->sum_thread / cpu_data->thread_count) / divisor); in timerlat_top_print() [all …]
|
| A D | osnoise_top.c | 68 struct osnoise_top_cpu *cpu_data; member 78 free(data->cpu_data); in osnoise_free_top() 96 data->cpu_data = calloc(1, sizeof(*data->cpu_data) * nr_cpus); in osnoise_alloc_top() 97 if (!data->cpu_data) in osnoise_alloc_top() 124 cpu_data = &data->cpu_data[cpu]; in osnoise_top_handler() 126 cpu_data->sum_cycles++; in osnoise_top_handler() 229 cpu_data = &data->cpu_data[cpu]; in osnoise_top_print() 231 if (!cpu_data->sum_runtime) in osnoise_top_print() 234 percentage = ((cpu_data->sum_runtime - cpu_data->sum_noise) * 10000000) in osnoise_top_print() 235 / cpu_data->sum_runtime; in osnoise_top_print() [all …]
|
| /linux/arch/sparc/kernel/ |
| A D | prom_64.c | 532 cpu_data(cpuid).clock_tick = in fill_in_one_cpu() 536 cpu_data(cpuid).dcache_size = in fill_in_one_cpu() 542 cpu_data(cpuid).icache_size = in fill_in_one_cpu() 548 cpu_data(cpuid).ecache_size = in fill_in_one_cpu() 554 cpu_data(cpuid).ecache_size = in fill_in_one_cpu() 564 cpu_data(cpuid).proc_id = portid; in fill_in_one_cpu() 566 cpu_data(cpuid).dcache_size = in fill_in_one_cpu() 571 cpu_data(cpuid).icache_size = in fill_in_one_cpu() 576 cpu_data(cpuid).ecache_size = in fill_in_one_cpu() 582 cpu_data(cpuid).core_id = 0; in fill_in_one_cpu() [all …]
|
| A D | smp_32.c | 61 cpu_data(id).udelay_val = loops_per_jiffy; in smp_store_cpu_info() 64 cpu_data(id).clock_tick = prom_getintdefault(cpu_node, in smp_store_cpu_info() 66 cpu_data(id).prom_node = cpu_node; in smp_store_cpu_info() 73 cpu_data(id).mid = mid; in smp_store_cpu_info() 83 bogosum += cpu_data(cpu).udelay_val; in smp_cpus_done() 377 cpu_data(i).udelay_val/(500000/HZ), in smp_bogo() 378 (cpu_data(i).udelay_val/(5000/HZ))%100); in smp_bogo()
|
| /linux/arch/mips/kernel/ |
| A D | proc.c | 40 unsigned int version = cpu_data[n].processor_id; in show_cpuinfo() 41 unsigned int fp_vers = cpu_data[n].fpu_id; in show_cpuinfo() 67 cpu_data[n].udelay_val / (500000/HZ), in show_cpuinfo() 68 (cpu_data[n].udelay_val / (5000/HZ)) % 100); in show_cpuinfo() 72 seq_printf(m, "tlb_entries\t\t: %d\n", cpu_data[n].tlbsize); in show_cpuinfo() 79 cpu_data[n].watch_reg_count); in show_cpuinfo() 80 for (i = 0; i < cpu_data[n].watch_reg_count; i++) in show_cpuinfo() 82 cpu_data[n].watch_reg_masks[i]); in show_cpuinfo() 281 cpu_data[n].srsets); in show_cpuinfo() 283 hweight8(cpu_data[n].kscratch_mask)); in show_cpuinfo() [all …]
|
| A D | smp-cps.c | 168 cpu_set_cluster(&cpu_data[nvpes + v], cl); in cps_smp_setup() 169 cpu_set_core(&cpu_data[nvpes + v], c); in cps_smp_setup() 170 cpu_set_vpe_id(&cpu_data[nvpes + v], v); in cps_smp_setup() 183 set_cpu_present(v, cpu_cluster(&cpu_data[v]) == 0); in cps_smp_setup() 387 unsigned core = cpu_core(&cpu_data[cpu]); in cps_boot_secondary() 388 unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]); in cps_boot_secondary() 503 core = cpu_core(&cpu_data[cpu]); in cps_shutdown_this_cpu() 506 vpe_id = cpu_vpe_id(&cpu_data[cpu]); in cps_shutdown_this_cpu() 602 unsigned vpe_id = cpu_vpe_id(&cpu_data[cpu]); in wait_for_sibling_halt() 618 unsigned core = cpu_core(&cpu_data[cpu]); in cps_cleanup_dead_cpu() [all …]
|
| /linux/arch/loongarch/kernel/ |
| A D | proc.c | 34 unsigned int isa = cpu_data[n].isa_level; in show_cpuinfo() 35 unsigned int version = cpu_data[n].processor_id & 0xff; in show_cpuinfo() 36 unsigned int fp_version = cpu_data[n].fpu_vers; in show_cpuinfo() 51 seq_printf(m, "package\t\t\t: %d\n", cpu_data[n].package); in show_cpuinfo() 52 seq_printf(m, "core\t\t\t: %d\n", cpu_data[n].core); in show_cpuinfo() 53 seq_printf(m, "global_id\t\t: %d\n", cpu_data[n].global_id); in show_cpuinfo() 63 seq_printf(m, "TLB Entries\t\t: %d\n", cpu_data[n].tlbsize); in show_cpuinfo() 98 cpu_data[n].watch_ireg_count, cpu_data[n].watch_dreg_count); in show_cpuinfo()
|
| /linux/arch/loongarch/include/asm/ |
| A D | cpu-features.h | 15 #define cpu_opt(opt) (cpu_data[0].options & (opt)) 16 #define cpu_has(feat) (cpu_data[0].options & BIT_ULL(feat)) 19 #define cpu_has_loongarch32 (cpu_data[0].isa_level & LOONGARCH_CPU_ISA_32BIT) 20 #define cpu_has_loongarch64 (cpu_data[0].isa_level & LOONGARCH_CPU_ISA_64BIT) 23 # define cpu_has_64bits (cpu_data[0].isa_level & LOONGARCH_CPU_ISA_64BIT) 30 # define cpu_vabits cpu_data[0].vabits 31 # define cpu_pabits cpu_data[0].pabits
|
| A D | cpu-info.h | 66 extern struct cpuinfo_loongarch cpu_data[]; 67 #define boot_cpu_data cpu_data[0] 68 #define current_cpu_data cpu_data[smp_processor_id()] 69 #define raw_current_cpu_data cpu_data[raw_smp_processor_id()] 101 struct cpuinfo_loongarch *infoa = &cpu_data[cpua]; in cpus_are_siblings() 102 struct cpuinfo_loongarch *infob = &cpu_data[cpub]; in cpus_are_siblings()
|
| A D | mmu_context.h | 26 return ~(u64)(cpu_asid_mask(&cpu_data[cpu])); in asid_version_mask() 31 return cpu_asid_mask(&cpu_data[cpu]) + 1; in asid_first_version() 35 #define asid_cache(cpu) (cpu_data[cpu].asid_cache) 36 #define cpu_asid(cpu, mm) (cpu_context((cpu), (mm)) & cpu_asid_mask(&cpu_data[cpu])) 56 if (!((++asid) & cpu_asid_mask(&cpu_data[cpu]))) in get_new_mmu_context()
|
| /linux/arch/x86/include/asm/ |
| A D | topology.h | 137 #define topology_logical_package_id(cpu) (cpu_data(cpu).topo.logical_pkg_id) 138 #define topology_physical_package_id(cpu) (cpu_data(cpu).topo.pkg_id) 139 #define topology_logical_die_id(cpu) (cpu_data(cpu).topo.logical_die_id) 140 #define topology_die_id(cpu) (cpu_data(cpu).topo.die_id) 141 #define topology_core_id(cpu) (cpu_data(cpu).topo.core_id) 142 #define topology_ppin(cpu) (cpu_data(cpu).ppin) 144 #define topology_amd_node_id(cpu) (cpu_data(cpu).topo.amd_node_id) 182 #define topology_cluster_id(cpu) (cpu_data(cpu).topo.l2c_id)
|
| /linux/kernel/trace/ |
| A D | trace_functions_graph.c | 31 struct fgraph_cpu_data __percpu *cpu_data; member 726 struct fgraph_cpu_data *cpu_data; in print_graph_entry_leaf() local 728 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_entry_leaf() 735 cpu_data->depth = call->depth - 1; in print_graph_entry_leaf() 778 struct fgraph_cpu_data *cpu_data; in print_graph_entry_nested() local 781 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_entry_nested() 782 cpu_data->depth = call->depth; in print_graph_entry_nested() 1015 struct fgraph_cpu_data *cpu_data; in print_graph_return() local 1018 cpu_data = per_cpu_ptr(data->cpu_data, cpu); in print_graph_return() 1324 if (!data->cpu_data) in graph_trace_open() [all …]
|
| /linux/arch/alpha/include/asm/ |
| A D | mmu_context.h | 91 #define cpu_last_asn(cpuid) (cpu_data[cpuid].last_asn) 143 cpu_data[cpu].asn_lock = 1; in ev5_switch_mm() 154 cpu_data[cpu].need_new_asn = 1; in ev5_switch_mm() 171 cpu_data[cpu].asn_lock = 0; \ 173 if (cpu_data[cpu].need_new_asn) { \ 175 cpu_data[cpu].need_new_asn = 0; \
|
| /linux/arch/mips/mm/ |
| A D | context.c | 36 if (!((asid += cpu_asid_inc()) & cpu_asid_mask(&cpu_data[cpu]))) { in get_new_mmu_context() 76 mmid = xchg_relaxed(&cpu_data[cpu].asid_cache, 0); in flush_context() 88 __set_bit(mmid & cpu_asid_mask(&cpu_data[cpu]), mmid_map); in flush_context() 210 old_active_mmid = READ_ONCE(cpu_data[cpu].asid_cache); in check_switch_mmu_context() 213 !cmpxchg_relaxed(&cpu_data[cpu].asid_cache, old_active_mmid, ctx)) { in check_switch_mmu_context() 220 WRITE_ONCE(cpu_data[cpu].asid_cache, ctx); in check_switch_mmu_context()
|
| /linux/arch/alpha/kernel/ |
| A D | smp.c | 55 struct cpuinfo_alpha cpu_data[NR_CPUS]; variable 56 EXPORT_SYMBOL(cpu_data); 83 cpu_data[cpuid].loops_per_jiffy = loops_per_jiffy; in smp_store_cpu_info() 84 cpu_data[cpuid].last_asn = ASN_FIRST_VERSION; in smp_store_cpu_info() 85 cpu_data[cpuid].need_new_asn = 0; in smp_store_cpu_info() 86 cpu_data[cpuid].asn_lock = 0; in smp_store_cpu_info() 95 cpu_data[cpuid].prof_counter = 1; in smp_setup_percpu_timer() 96 cpu_data[cpuid].prof_multiplier = 1; in smp_setup_percpu_timer() 487 bogosum += cpu_data[cpu].loops_per_jiffy; in smp_cpus_done() 554 cpu_data[this_cpu].ipi_count++; in handle_ipi() [all …]
|
| /linux/arch/mips/loongson64/ |
| A D | smp.c | 417 cpu_set_core(&cpu_data[cpu], in loongson3_init_secondary() 419 cpu_data[cpu].package = in loongson3_init_secondary() 466 cpu_set_cluster(&cpu_data[num], i / 4); in loongson3_smp_setup() 486 cpu_set_core(&cpu_data[0], in loongson3_smp_setup() 571 [sets] "r" (cpu_data[smp_processor_id()].dcache.sets)); in loongson3_type1_play_dead() 633 [sets] "r" (cpu_data[smp_processor_id()].dcache.sets)); in loongson3_type2_play_dead() 717 [sets] "r" (cpu_data[smp_processor_id()].dcache.sets), in loongson3_type3_play_dead() 800 uint64_t core_id = cpu_core(&cpu_data[cpu]); in loongson3_disable_clock() 801 uint64_t package_id = cpu_data[cpu].package; in loongson3_disable_clock() 817 uint64_t core_id = cpu_core(&cpu_data[cpu]); in loongson3_enable_clock() [all …]
|
| /linux/arch/sh/include/asm/ |
| A D | processor.h | 89 extern struct sh_cpuinfo cpu_data[]; 90 #define boot_cpu_data cpu_data[0] 91 #define current_cpu_data cpu_data[smp_processor_id()] 92 #define raw_current_cpu_data cpu_data[raw_smp_processor_id()]
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| A D | perf_buffer.c | 19 int cpu_data = *(int *)data, duration = 0; in on_sample() local 22 if (cpu_data != cpu) in on_sample() 23 CHECK(cpu_data != cpu, "check_cpu_data", in on_sample() 24 "cpu_data %d != cpu %d\n", cpu_data, cpu); in on_sample()
|
| /linux/arch/x86/kernel/ |
| A D | smpboot.c | 323 struct cpuinfo_x86 *c = &cpu_data(id); in smp_store_cpu_info() 557 struct cpuinfo_x86 *c = &cpu_data(cpu); in set_cpu_sibling_map() 574 o = &cpu_data(i); in set_cpu_sibling_map() 597 cpu_data(i).smt_active = threads > 1; in set_cpu_sibling_map() 604 o = &cpu_data(i); in set_cpu_sibling_map() 625 cpu_data(i).booted_cores++; in set_cpu_sibling_map() 653 bogosum += cpu_data(cpu).loops_per_jiffy; in impress_friends() 1106 print_cpu_info(&cpu_data(0)); in native_smp_prepare_cpus() 1177 struct cpuinfo_x86 *c = &cpu_data(cpu); in remove_siblinginfo() 1185 cpu_data(sibling).booted_cores--; in remove_siblinginfo() [all …]
|