Searched refs:X86_PMC_IDX_MAX (Results 1 – 11 of 11) sorted by relevance
| /linux/arch/x86/events/ |
| A D | perf_event.h | 51 unsigned long idxmsk[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; 114 struct perf_event *owners[X86_PMC_IDX_MAX]; 115 struct event_constraint event_constraints[X86_PMC_IDX_MAX]; 189 enum intel_excl_state_type state[X86_PMC_IDX_MAX]; 231 struct perf_event *events[X86_PMC_IDX_MAX]; /* in counter order */ 232 unsigned long active_mask[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; 233 unsigned long dirty[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; 243 int assign[X86_PMC_IDX_MAX]; /* event to counter assignment */ 244 u64 tags[X86_PMC_IDX_MAX]; 247 struct event_constraint *event_constraint[X86_PMC_IDX_MAX]; [all …]
|
| A D | core.c | 874 for_each_set_bit_from(idx, c->idxmsk, X86_PMC_IDX_MAX) { in __perf_sched_find_counter() 996 for (i = 0, wmin = X86_PMC_IDX_MAX, wmax = 0; i < n; i++) { in x86_schedule_events() 1353 static DEFINE_PER_CPU(u64 [X86_PMC_IDX_MAX], pmc_prev_left); 1450 int assign[X86_PMC_IDX_MAX]; in x86_pmu_add() 2286 int assign[X86_PMC_IDX_MAX]; in x86_pmu_commit_txn() 2493 if (bitmap_empty(cpuc->dirty, X86_PMC_IDX_MAX)) in perf_clear_dirty_counters() 2496 for_each_set_bit(i, cpuc->dirty, X86_PMC_IDX_MAX) { in perf_clear_dirty_counters() 2508 bitmap_zero(cpuc->dirty, X86_PMC_IDX_MAX); in perf_clear_dirty_counters()
|
| /linux/arch/x86/kvm/ |
| A D | pmu.c | 301 for_each_set_bit(bit, pmu->reprogram_pmi, X86_PMC_IDX_MAX) { in kvm_pmu_handle_event() 467 DECLARE_BITMAP(bitmask, X86_PMC_IDX_MAX); in kvm_pmu_cleanup() 473 pmu->pmc_in_use, X86_PMC_IDX_MAX); in kvm_pmu_cleanup() 475 for_each_set_bit(i, bitmask, X86_PMC_IDX_MAX) { in kvm_pmu_cleanup() 485 bitmap_zero(pmu->pmc_in_use, X86_PMC_IDX_MAX); in kvm_pmu_cleanup()
|
| /linux/arch/x86/include/asm/ |
| A D | perf_event.h | 13 #define X86_PMC_IDX_MAX 64 macro
|
| A D | kvm_host.h | 510 DECLARE_BITMAP(reprogram_pmi, X86_PMC_IDX_MAX); 511 DECLARE_BITMAP(all_valid_pmc_idx, X86_PMC_IDX_MAX); 512 DECLARE_BITMAP(pmc_in_use, X86_PMC_IDX_MAX);
|
| /linux/arch/x86/events/intel/ |
| A D | knc.c | 242 for_each_set_bit(bit, (unsigned long *)&status, X86_PMC_IDX_MAX) { in knc_pmu_handle_irq()
|
| A D | p4.c | 986 static DEFINE_PER_CPU(unsigned long [BITS_TO_LONGS(X86_PMC_IDX_MAX)], p4_running); 1219 unsigned long used_mask[BITS_TO_LONGS(X86_PMC_IDX_MAX)]; in p4_pmu_schedule_events() 1229 bitmap_zero(used_mask, X86_PMC_IDX_MAX); in p4_pmu_schedule_events()
|
| A D | core.c | 2927 for_each_set_bit(bit, (unsigned long *)&status, X86_PMC_IDX_MAX) { in handle_pmi_common() 3446 for_each_set_bit(i, c->idxmsk, X86_PMC_IDX_MAX) { in intel_get_excl_constraints() 3498 bitmap_copy(c1->idxmsk, c2->idxmsk, X86_PMC_IDX_MAX); in intel_get_event_constraints() 4306 size_t sz = X86_PMC_IDX_MAX * sizeof(struct event_constraint); in intel_cpuc_prepare()
|
| /linux/arch/x86/events/amd/ |
| A D | core.c | 271 static unsigned int event_offsets[X86_PMC_IDX_MAX] __read_mostly; 272 static unsigned int count_offsets[X86_PMC_IDX_MAX] __read_mostly;
|
| /linux/arch/x86/events/zhaoxin/ |
| A D | core.c | 386 for_each_set_bit(bit, (unsigned long *)&status, X86_PMC_IDX_MAX) { in zhaoxin_pmu_handle_irq()
|
| /linux/arch/x86/kvm/vmx/ |
| A D | pmu_intel.c | 67 for_each_set_bit(bit, (unsigned long *)&diff, X86_PMC_IDX_MAX) in global_ctrl_changed()
|
Completed in 42 milliseconds