| /linux/arch/x86/xen/ |
| A D | pmu.c | 143 static int is_intel_pmu_msr(u32 msr_index, int *type, int *index) in is_intel_pmu_msr() argument 147 switch (msr_index) { in is_intel_pmu_msr() 162 if ((msr_index >= MSR_CORE_PERF_FIXED_CTR0) && in is_intel_pmu_msr() 163 (msr_index < MSR_CORE_PERF_FIXED_CTR0 + in is_intel_pmu_msr() 165 *index = msr_index - MSR_CORE_PERF_FIXED_CTR0; in is_intel_pmu_msr() 170 if ((msr_index >= MSR_P6_EVNTSEL0) && in is_intel_pmu_msr() 171 (msr_index < MSR_P6_EVNTSEL0 + intel_num_arch_counters)) { in is_intel_pmu_msr() 172 *index = msr_index - MSR_P6_EVNTSEL0; in is_intel_pmu_msr() 177 msr_index_pmc = msr_index & MSR_PMC_ALIAS_MASK; in is_intel_pmu_msr()
|
| /linux/tools/testing/selftests/kvm/x86_64/ |
| A D | userspace_msr_exit_test.c | 418 static void process_rdmsr(struct kvm_vm *vm, uint32_t msr_index) in process_rdmsr() argument 428 TEST_ASSERT(run->msr.index == msr_index, in process_rdmsr() 430 run->msr.index, msr_index); in process_rdmsr() 453 static void process_wrmsr(struct kvm_vm *vm, uint32_t msr_index) in process_wrmsr() argument 463 TEST_ASSERT(run->msr.index == msr_index, in process_wrmsr() 465 run->msr.index, msr_index); in process_wrmsr() 529 static void run_guest_then_process_rdmsr(struct kvm_vm *vm, uint32_t msr_index) in run_guest_then_process_rdmsr() argument 532 process_rdmsr(vm, msr_index); in run_guest_then_process_rdmsr() 535 static void run_guest_then_process_wrmsr(struct kvm_vm *vm, uint32_t msr_index) in run_guest_then_process_wrmsr() argument 538 process_wrmsr(vm, msr_index); in run_guest_then_process_wrmsr()
|
| A D | xss_msr_test.c | 20 bool is_supported_msr(u32 msr_index) in is_supported_msr() argument 28 if (list->indices[i] == msr_index) { in is_supported_msr()
|
| /linux/drivers/thermal/intel/ |
| A D | intel_powerclamp.c | 207 int msr_index; member 212 .msr_index = MSR_PKG_C##id##_RESIDENCY, \ 233 while (info->msr_index) { in has_pkg_state_counter() 234 if (!rdmsrl_safe(info->msr_index, &val)) in has_pkg_state_counter() 248 while (info->msr_index) { in pkg_state_counter() 250 if (!rdmsrl_safe(info->msr_index, &val)) in pkg_state_counter()
|
| /linux/tools/testing/selftests/kvm/include/x86_64/ |
| A D | processor.h | 357 uint64_t kvm_get_feature_msr(uint64_t msr_index); 373 uint64_t vcpu_get_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index); 374 int _vcpu_set_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index, 376 void vcpu_set_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index,
|
| /linux/arch/x86/kvm/vmx/ |
| A D | nested.c | 944 u32 msr_index, in nested_vmx_get_vmexit_msr_value() argument 968 msr_index); in nested_vmx_get_vmexit_msr_value() 1036 if (e.index == msr_index) in nested_msr_store_list_has_msr() 1043 u32 msr_index) in prepare_vmx_msr_autostore_list() argument 1067 msr_index); in prepare_vmx_msr_autostore_list() 1231 switch (msr_index) { in vmx_restore_control_msr() 1328 switch (msr_index) { in vmx_restore_fixed0_msr() 1366 switch (msr_index) { in vmx_set_vmx_msr() 1422 switch (msr_index) { in vmx_get_vmx_msr() 5696 msr_index -= 0xc0000000; in nested_vmx_exit_handled_msr() [all …]
|
| A D | nested.h | 31 int vmx_set_vmx_msr(struct kvm_vcpu *vcpu, u32 msr_index, u64 data); 32 int vmx_get_vmx_msr(struct nested_vmx_msrs *msrs, u32 msr_index, u64 *pdata);
|
| A D | evmcs.c | 344 void nested_evmcs_filter_control_msr(u32 msr_index, u64 *pdata) in nested_evmcs_filter_control_msr() argument 353 switch (msr_index) { in nested_evmcs_filter_control_msr()
|
| A D | evmcs.h | 215 void nested_evmcs_filter_control_msr(u32 msr_index, u64 *pdata);
|
| A D | vmx.c | 1943 u32 msr_index = msr_info->index; in vmx_set_msr() local 1947 switch (msr_index) { in vmx_set_msr() 2144 [msr_index - MSR_IA32_SGXLEPUBKEYHASH0] = data; in vmx_set_msr() 2151 return vmx_set_vmx_msr(vcpu, msr_index, data); in vmx_set_msr() 2226 msr = vmx_find_uret_msr(vmx, msr_index); in vmx_set_msr()
|
| /linux/tools/testing/selftests/kvm/lib/x86_64/ |
| A D | processor.c | 755 uint64_t kvm_get_feature_msr(uint64_t msr_index) in kvm_get_feature_msr() argument 764 buffer.entry.index = msr_index; in kvm_get_feature_msr() 892 uint64_t vcpu_get_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index) in vcpu_get_msr() argument 903 buffer.entry.index = msr_index; in vcpu_get_msr() 926 int _vcpu_set_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index, in _vcpu_set_msr() argument 939 buffer.entry.index = msr_index; in _vcpu_set_msr() 960 void vcpu_set_msr(struct kvm_vm *vm, uint32_t vcpuid, uint64_t msr_index, in vcpu_set_msr() argument 965 r = _vcpu_set_msr(vm, vcpuid, msr_index, msr_value); in vcpu_set_msr()
|
| /linux/arch/x86/kvm/ |
| A D | kvm_emulate.h | 213 int (*set_msr)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 data); 214 int (*get_msr)(struct x86_emulate_ctxt *ctxt, u32 msr_index, u64 *pdata);
|
| A D | emulate.c | 3609 u64 msr_index = reg_read(ctxt, VCPU_REGS_RCX); in em_wrmsr() local 3615 r = ctxt->ops->set_msr(ctxt, msr_index, msr_data); in em_wrmsr() 3628 u64 msr_index = reg_read(ctxt, VCPU_REGS_RCX); in em_rdmsr() local 3632 r = ctxt->ops->get_msr(ctxt, msr_index, &msr_data); in em_rdmsr()
|
| A D | x86.c | 7390 u32 msr_index, u64 *pdata) in emulator_get_msr() argument 7395 r = kvm_get_msr(vcpu, msr_index, pdata); in emulator_get_msr() 7397 if (r && kvm_get_msr_user_space(vcpu, msr_index, r)) { in emulator_get_msr() 7406 u32 msr_index, u64 data) in emulator_set_msr() argument 7411 r = kvm_set_msr(vcpu, msr_index, data); in emulator_set_msr() 7413 if (r && kvm_set_msr_user_space(vcpu, msr_index, data, r)) { in emulator_set_msr()
|