Searched refs:counts (Results 1 – 8 of 8) sorted by relevance
| /arch/x86/kernel/cpu/resctrl/ |
| A D | pseudo_lock.c | 418 counts->miss_before = miss_before; in measure_residency_fn() 419 counts->hits_before = hits_before; in measure_residency_fn() 420 counts->miss_after = miss_after; in measure_residency_fn() 421 counts->hits_after = hits_after; in measure_residency_fn() 455 trace_pseudo_lock_l2(counts.hits_after - counts.hits_before, in resctrl_arch_measure_l2_residency() 456 counts.miss_after - counts.miss_before); in resctrl_arch_measure_l2_residency() 495 counts.miss_after -= counts.miss_before; in resctrl_arch_measure_l3_residency() 505 counts.hits_after -= counts.hits_before; in resctrl_arch_measure_l3_residency() 507 counts.hits_after -= min(counts.miss_after, counts.hits_after); in resctrl_arch_measure_l3_residency() 509 counts.hits_after -= counts.hits_before; in resctrl_arch_measure_l3_residency() [all …]
|
| /arch/powerpc/platforms/pseries/ |
| A D | msi.c | 235 counts->num_devices++; in count_non_bridge_devices() 247 req = counts->request; in count_spare_msis() 261 if (req < counts->quota) in count_spare_msis() 262 counts->spare += counts->quota - req; in count_spare_msis() 264 counts->over_quota++; in count_spare_msis() 272 struct msi_counts counts; in msi_quota_for_device() local 300 counts.quota = total / counts.num_devices; in msi_quota_for_device() 306 counts.request = request; in msi_quota_for_device() 311 counts.spare += total % counts.num_devices; in msi_quota_for_device() 314 if (counts.over_quota) in msi_quota_for_device() [all …]
|
| /arch/x86/kernel/cpu/ |
| A D | aperfmperf.c | 190 u64 ratios, counts; in skx_set_max_freq_ratio() local 204 err = rdmsrq_safe(MSR_TURBO_RATIO_LIMIT1, &counts); in skx_set_max_freq_ratio() 209 group_size = (counts >> i) & 0xFF; in skx_set_max_freq_ratio()
|
| /arch/x86/ras/ |
| A D | Kconfig | 7 page PFN and counts their repeated occurrence. Once the counter for a
|
| /arch/x86/events/intel/ |
| A D | ds.c | 2494 short counts[INTEL_PMC_IDX_FIXED + MAX_FIXED_PEBS_EVENTS] = {}; in intel_pmu_drain_pebs_nhm() local 2530 counts[bit]++; in intel_pmu_drain_pebs_nhm() 2574 counts[bit]++; in intel_pmu_drain_pebs_nhm() 2578 if ((counts[bit] == 0) && (error[bit] == 0)) in intel_pmu_drain_pebs_nhm() 2596 if (counts[bit]) { in intel_pmu_drain_pebs_nhm() 2598 top, bit, counts[bit], in intel_pmu_drain_pebs_nhm() 2606 short counts[INTEL_PMC_IDX_FIXED + MAX_FIXED_PEBS_EVENTS] = {}; in intel_pmu_drain_pebs_icl() local 2653 if (counts[bit]++) { in intel_pmu_drain_pebs_icl() 2662 if (!counts[bit]) in intel_pmu_drain_pebs_icl() 2668 counts[bit], setup_pebs_adaptive_sample_data); in intel_pmu_drain_pebs_icl()
|
| /arch/sparc/lib/ |
| A D | M7memset.S | 129 cmp %o2, 7 ! if small counts, just write bytes
|
| /arch/sh/lib/ |
| A D | memcpy-sh4.S | 197 ! cycle counts for differnet sizes using byte-at-a-time vs. optimised):
|
| /arch/ |
| A D | Kconfig | 1575 bool "Locking event counts collection" 1581 differences. The counts are reported via debugfs.
|
Completed in 22 milliseconds