Searched refs:x86_pmu (Results 1 – 16 of 16) sorted by relevance
| /linux/arch/x86/events/intel/ |
| A D | lbr.c | 261 if (!x86_pmu.lbr_nr) in intel_pmu_lbr_reset() 264 x86_pmu.lbr_reset(); in intel_pmu_lbr_reset() 555 if (i < x86_pmu.lbr_nr) in intel_pmu_arch_lbr_save() 579 x86_pmu.lbr_save(ctx); in __intel_pmu_lbr_save() 651 if (!x86_pmu.lbr_nr) in intel_pmu_lbr_add() 732 if (!x86_pmu.lbr_nr) in intel_pmu_lbr_del() 1136 if (!x86_pmu.lbr_nr) in intel_pmu_setup_lbr_filter() 1585 x86_pmu.lbr_nr = 16; in intel_pmu_lbr_init_snb() 1612 x86_pmu.lbr_nr = 16; in intel_pmu_lbr_init_hsw() 1631 x86_pmu.lbr_nr = 32; in intel_pmu_lbr_init_skl() [all …]
|
| A D | core.c | 2826 if (x86_pmu.lbr_nr) { in intel_pmu_reset() 3930 if (x86_pmu.pebs && x86_pmu.pebs_no_isolation) { in intel_guest_get_msrs() 4299 if (is_hybrid() || x86_pmu.extra_regs || x86_pmu.lbr_sel_map) { in intel_cpuc_prepare() 4733 x86_pmu.pebs = 0; in intel_clovertown_quirk() 4812 unsigned long lbr_from_nr = x86_pmu.lbr_from + x86_pmu.lbr_nr; in is_lbr_from() 5605 x86_pmu = core_pmu; in intel_pmu_init() 5607 x86_pmu = intel_pmu; in intel_pmu_init() 6337 if (x86_pmu.lbr_tos && !check_msr(x86_pmu.lbr_tos, 0x3UL)) in intel_pmu_init() 6338 x86_pmu.lbr_nr = 0; in intel_pmu_init() 6342 x86_pmu.lbr_nr = 0; in intel_pmu_init() [all …]
|
| A D | ds.c | 408 if (!x86_pmu.pebs) in alloc_pebs_buffer() 433 max = x86_pmu.pebs_record_size * (bsiz / x86_pmu.pebs_record_size); in alloc_pebs_buffer() 443 if (!x86_pmu.pebs) in release_pebs_buffer() 463 if (!x86_pmu.bts) in alloc_bts_buffer() 490 if (!x86_pmu.bts) in release_bts_buffer() 518 if (!x86_pmu.bts && !x86_pmu.pebs) in release_ds_buffers() 547 if (!x86_pmu.bts && !x86_pmu.pebs) in reserve_ds_buffers() 550 if (!x86_pmu.bts) in reserve_ds_buffers() 553 if (!x86_pmu.pebs) in reserve_ds_buffers() 2163 if (x86_pmu.pebs) { in intel_ds_init() [all …]
|
| A D | p6.c | 201 static __initconst const struct x86_pmu p6_pmu = { 242 x86_pmu.attr_rdpmc_broken = 1; in p6_pmu_rdpmc_quirk() 243 x86_pmu.attr_rdpmc = 0; in p6_pmu_rdpmc_quirk() 249 x86_pmu = p6_pmu; in p6_pmu_init()
|
| A D | knc.c | 290 static const struct x86_pmu knc_pmu __initconst = { 316 x86_pmu = knc_pmu; in knc_pmu_init()
|
| A D | p4.c | 922 for (idx = 0; idx < x86_pmu.num_counters; idx++) { in p4_pmu_disable_all() 1001 for (idx = 0; idx < x86_pmu.num_counters; idx++) { in p4_pmu_enable_all() 1020 for (idx = 0; idx < x86_pmu.num_counters; idx++) { in p4_pmu_handle_irq() 1039 if (!overflow && (val & (1ULL << (x86_pmu.cntval_bits - 1)))) in p4_pmu_handle_irq() 1312 static __initconst const struct x86_pmu p4_pmu = { 1370 x86_pmu = p4_pmu; in p4_pmu_init() 1381 for (i = 0; i < x86_pmu.num_counters; i++) { in p4_pmu_init()
|
| A D | bts.c | 590 if (!boot_cpu_has(X86_FEATURE_DTES64) || !x86_pmu.bts) in bts_init()
|
| /linux/arch/x86/events/ |
| A D | core.c | 47 struct x86_pmu x86_pmu __read_mostly; 547 if (x86_pmu.pebs_active && !x86_pmu.pebs_broken) { in x86_pmu_max_precise() 551 if (x86_pmu.lbr_nr || x86_pmu.intel_cap.pebs_format >= 2) in x86_pmu_max_precise() 1560 if (x86_pmu.lbr_nr) { in perf_event_print_debug() 1774 if (x86_pmu.cpu_dead) in x86_pmu_dead_cpu() 1775 x86_pmu.cpu_dead(cpu); in x86_pmu_dead_cpu() 1800 if (x86_pmu.cpu_dying) in x86_pmu_dying_cpu() 1810 x86_pmu.apic = 0; in pmu_check_apic() 2115 x86_pmu.intel_ctrl = (1 << x86_pmu.num_counters) - 1; in init_hw_perf_events() 2137 if (!x86_pmu.read) in init_hw_perf_events() [all …]
|
| A D | perf_event.h | 676 typeof(&x86_pmu._field) __Fp = &x86_pmu._field; \ 696 bool __Fp = x86_pmu._field; \ 719 struct x86_pmu { struct 964 __quirk.next = x86_pmu.quirks; \ 965 x86_pmu.quirks = &__quirk; \ 1023 extern struct x86_pmu x86_pmu __read_mostly; 1035 return x86_pmu.lbr_sel_map && in x86_pmu_has_lbr_callstack() 1066 return x86_pmu.eventsel + (x86_pmu.addr_offset ? in x86_pmu_config_addr() 1072 return x86_pmu.perfctr + (x86_pmu.addr_offset ? in x86_pmu_event_addr() 1078 return x86_pmu.rdpmc_index ? x86_pmu.rdpmc_index(index) : index; in x86_pmu_rdpmc_index() [all …]
|
| /linux/arch/x86/events/zhaoxin/ |
| A D | core.c | 370 if (x86_pmu.enabled_ack) in zhaoxin_pmu_handle_irq() 427 if (x86_pmu.event_constraints) { in zhaoxin_get_event_constraints() 529 x86_pmu = zhaoxin_pmu; in zhaoxin_pmu_init() 532 x86_pmu.version = version; in zhaoxin_pmu_init() 536 x86_pmu.events_maskl = ebx.full; in zhaoxin_pmu_init() 546 x86_pmu.max_period = x86_pmu.cntval_mask >> 1; in zhaoxin_pmu_init() 549 x86_pmu.enabled_ack = 1; in zhaoxin_pmu_init() 601 x86_pmu.intel_ctrl = (1 << (x86_pmu.num_counters)) - 1; in zhaoxin_pmu_init() 602 x86_pmu.intel_ctrl |= ((1LL << x86_pmu.num_counters_fixed)-1) << INTEL_PMC_IDX_FIXED; in zhaoxin_pmu_init() 604 if (x86_pmu.event_constraints) { in zhaoxin_pmu_init() [all …]
|
| /linux/arch/x86/events/amd/ |
| A D | core.c | 319 if (!(x86_pmu.flags & PMU_FL_PAIR)) in amd_is_pair_event_code() 519 if (!x86_pmu.amd_nb_constraints) in amd_pmu_cpu_prepare() 538 if (!x86_pmu.amd_nb_constraints) in amd_pmu_cpu_starting() 564 if (!x86_pmu.amd_nb_constraints) in amd_pmu_cpu_dead() 957 x86_pmu.eventsel = MSR_F15H_PERF_CTL; in amd_core_pmu_init() 958 x86_pmu.perfctr = MSR_F15H_PERF_CTR; in amd_core_pmu_init() 964 x86_pmu.amd_nb_constraints = 0; in amd_core_pmu_init() 983 x86_pmu.num_counters / 2, 0, in amd_core_pmu_init() 989 x86_pmu.flags |= PMU_FL_PAIR; in amd_core_pmu_init() 1004 x86_pmu = amd_pmu; in amd_pmu_init() [all …]
|
| /linux/arch/x86/kvm/vmx/ |
| A D | pmu_intel.c | 467 struct x86_pmu_capability x86_pmu; in intel_pmu_refresh() local 489 perf_get_x86_pmu_capability(&x86_pmu); in intel_pmu_refresh() 492 x86_pmu.num_counters_gp); in intel_pmu_refresh() 493 eax.split.bit_width = min_t(int, eax.split.bit_width, x86_pmu.bit_width_gp); in intel_pmu_refresh() 495 eax.split.mask_length = min_t(int, eax.split.mask_length, x86_pmu.events_mask_len); in intel_pmu_refresh() 504 x86_pmu.num_counters_fixed); in intel_pmu_refresh() 506 edx.split.bit_width_fixed, x86_pmu.bit_width_fixed); in intel_pmu_refresh()
|
| /linux/Documentation/translations/zh_CN/riscv/ |
| A D | pmu.rst | 84 hwc->sample_period = x86_pmu.max_period;
|
| /linux/arch/x86/xen/ |
| A D | pmu.c | 501 if (x86_pmu.handle_irq(®s)) in xen_pmu_irq_handler()
|
| /linux/Documentation/riscv/ |
| A D | pmu.rst | 85 hwc->sample_period = x86_pmu.max_period;
|
| /linux/arch/x86/kvm/ |
| A D | x86.c | 6357 struct x86_pmu_capability x86_pmu; in kvm_init_msr_list() local 6364 perf_get_x86_pmu_capability(&x86_pmu); in kvm_init_msr_list() 6417 min(INTEL_PMC_MAX_GENERIC, x86_pmu.num_counters_gp)) in kvm_init_msr_list() 6422 min(INTEL_PMC_MAX_GENERIC, x86_pmu.num_counters_gp)) in kvm_init_msr_list()
|
Completed in 69 milliseconds