Searched refs:vcpu_to_pmu (Results 1 – 10 of 10) sorted by relevance
91 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_is_valid_rdpmc_ecx()102 return amd_pmc_idx_to_pmc(vcpu_to_pmu(vcpu), idx & ~(3u << 30)); in amd_rdpmc_ecx_to_pmc()113 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_msr_idx_to_pmc()124 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_get_msr()146 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_set_msr()174 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_refresh()194 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_init()210 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in amd_pmu_reset()
123 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_is_valid_rdpmc_ecx()135 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_rdpmc_ecx_to_pmc()195 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_is_valid_msr()230 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_msr_idx_to_pmc()247 vcpu_to_pmu(vcpu)->event_count--; in intel_pmu_release_guest_lbr_event()254 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_create_guest_lbr_event()347 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_get_msr()397 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_set_msr()512 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_refresh()622 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in intel_pmu_init()[all …]
2646 intel_pmu_has_perf_global_ctrl(vcpu_to_pmu(vcpu)) && in prepare_vmcs02()2922 CC(!kvm_valid_perf_global_ctrl(vcpu_to_pmu(vcpu), in nested_vmx_check_host_state()3047 CC(!kvm_valid_perf_global_ctrl(vcpu_to_pmu(vcpu), in nested_vmx_check_guest_state()4516 intel_pmu_has_perf_global_ctrl(vcpu_to_pmu(vcpu))) in load_vmcs12_host_state()
2435 if (data && !vcpu_to_pmu(vcpu)->version) in vmx_set_msr()7086 struct kvm_pmu *pmu = vcpu_to_pmu(&vmx->vcpu); in atomic_switch_perf_msrs()
202 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in pmu_ctr_read()265 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_incr_fw()282 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_read_hpm()316 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_num_ctrs()326 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_info()342 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_start()398 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_stop()465 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_ctr_cfg_match()548 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_init()613 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_riscv_vcpu_pmu_deinit()
21 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_sbi_ext_pmu_handler()76 struct kvm_pmu *kvpmu = vcpu_to_pmu(vcpu); in kvm_sbi_ext_pmu_probe()
463 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_handle_event()529 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_rdpmc()568 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_mark_pmc_in_use()597 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_reset()605 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_init()618 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_cleanup()680 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_pmu_trigger_event()830 atomic64_set(&vcpu_to_pmu(vcpu)->__reprogram_pmi, -1ull); in kvm_vm_ioctl_set_pmu_event_filter()
7 #define vcpu_to_pmu(vcpu) (&(vcpu)->arch.pmu) macro
12155 struct kvm_pmu *pmu = vcpu_to_pmu(vcpu); in kvm_arch_sched_in()
55 #define vcpu_to_pmu(vcpu) (&(vcpu)->arch.pmu_context) macro
Completed in 59 milliseconds