| /tools/testing/selftests/kvm/x86/ |
| A D | state_test.c | 39 GUEST_ASSERT(svm->vmcb_gpa); in svm_l1_guest_code() 93 GUEST_ASSERT(!vmlaunch()); in vmx_l1_guest_code() 98 GUEST_ASSERT(vmlaunch()); in vmx_l1_guest_code() 100 GUEST_ASSERT(!vmresume()); in vmx_l1_guest_code() 115 GUEST_ASSERT(vmlaunch()); in vmx_l1_guest_code() 117 GUEST_ASSERT(vmlaunch()); in vmx_l1_guest_code() 118 GUEST_ASSERT(vmresume()); in vmx_l1_guest_code() 130 GUEST_ASSERT(vmlaunch()); in vmx_l1_guest_code() 131 GUEST_ASSERT(vmresume()); in vmx_l1_guest_code() 134 GUEST_ASSERT(vmlaunch()); in vmx_l1_guest_code() [all …]
|
| A D | hyperv_evmcs.c | 109 GUEST_ASSERT(vmlaunch()); in guest_code() 124 GUEST_ASSERT(!vmlaunch()); in guest_code() 134 GUEST_ASSERT(!vmresume()); in guest_code() 145 GUEST_ASSERT(!vmresume()); in guest_code() 151 GUEST_ASSERT(!vmresume()); in guest_code() 159 GUEST_ASSERT(!vmresume()); in guest_code() 166 GUEST_ASSERT(!vmresume()); in guest_code() 174 GUEST_ASSERT(!vmresume()); in guest_code() 179 GUEST_ASSERT(!vmresume()); in guest_code() 182 GUEST_ASSERT(!vmresume()); in guest_code() [all …]
|
| A D | amx_test.c | 95 GUEST_ASSERT(xtile.xsave_offset == 2816); in check_xtile_info() 97 GUEST_ASSERT(xtile.xsave_size == 8192); in check_xtile_info() 98 GUEST_ASSERT(sizeof(struct tile_data) >= xtile.xsave_size); in check_xtile_info() 104 GUEST_ASSERT(this_cpu_has_p(X86_PROPERTY_AMX_NR_TILE_REGS)); in check_xtile_info() 106 GUEST_ASSERT(xtile.max_names == 8); in check_xtile_info() 108 GUEST_ASSERT(xtile.bytes_per_tile == 1024); in check_xtile_info() 110 GUEST_ASSERT(xtile.bytes_per_row == 64); in check_xtile_info() 112 GUEST_ASSERT(xtile.max_rows == 16); in check_xtile_info() 131 GUEST_ASSERT(this_cpu_has(X86_FEATURE_XSAVE) && in guest_code() 139 GUEST_ASSERT(rdmsr(MSR_IA32_XFD) == 0); in guest_code() [all …]
|
| A D | hyperv_ipi.c | 107 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); in sender_guest_code() 108 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]); in sender_guest_code() 114 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); in sender_guest_code() 115 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]); in sender_guest_code() 127 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); in sender_guest_code() 128 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]); in sender_guest_code() 136 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ++ipis_expected[0]); in sender_guest_code() 137 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ipis_expected[1]); in sender_guest_code() 149 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_1] == ipis_expected[0]); in sender_guest_code() 150 GUEST_ASSERT(ipis_rcvd[RECEIVER_VCPU_ID_2] == ++ipis_expected[1]); in sender_guest_code() [all …]
|
| A D | hyperv_clock.c | 57 GUEST_ASSERT(tsc_freq > 0); in check_tsc_msr_rdtsc() 68 GUEST_ASSERT(r2 > r1 && t2 > t1); in check_tsc_msr_rdtsc() 93 GUEST_ASSERT(r1 >= t1 && r1 - t1 < 100000); in check_tsc_msr_tsc_page() 98 GUEST_ASSERT(r2 >= t1 && r2 - t2 < 100000); in check_tsc_msr_tsc_page() 116 GUEST_ASSERT(tsc_page->tsc_sequence == 0); in guest_main() 117 GUEST_ASSERT(tsc_page->tsc_scale == 0); in guest_main() 118 GUEST_ASSERT(tsc_page->tsc_offset == 0); in guest_main() 124 GUEST_ASSERT(tsc_page->tsc_sequence != 0); in guest_main() 169 GUEST_ASSERT(tsc_page->tsc_sequence == 0); in guest_main() 170 GUEST_ASSERT(tsc_page->tsc_offset == 0); in guest_main() [all …]
|
| A D | vmx_tsc_adjust_test.c | 61 GUEST_ASSERT(adjust <= max); in check_ia32_tsc_adjust() 82 GUEST_ASSERT(rdtsc() < TSC_ADJUST_VALUE); in l1_guest_code() 86 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code() 87 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code() 100 GUEST_ASSERT(!vmlaunch()); in l1_guest_code() 101 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == in l1_guest_code() 106 GUEST_ASSERT(!vmlaunch()); in l1_guest_code() 107 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == EXIT_REASON_VMCALL); in l1_guest_code()
|
| A D | vmx_preemption_timer_test.c | 75 GUEST_ASSERT(vmx_pages->vmcs_gpa); in l1_guest_code() 76 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code() 77 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code() 78 GUEST_ASSERT(vmptrstz() == vmx_pages->vmcs_gpa); in l1_guest_code() 96 GUEST_ASSERT(!vmlaunch()); in l1_guest_code() 97 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == EXIT_REASON_VMCALL); in l1_guest_code() 103 GUEST_ASSERT(!vmwrite(PIN_BASED_VM_EXEC_CONTROL, in l1_guest_code() 107 GUEST_ASSERT(!vmwrite(VMX_PREEMPTION_TIMER_VALUE, in l1_guest_code() 116 GUEST_ASSERT(!vmresume()); in l1_guest_code() 124 GUEST_ASSERT(l2_save_restore_done); in l1_guest_code() [all …]
|
| A D | userspace_msr_exit_test.c | 220 GUEST_ASSERT(data == 0); in guest_code_filter_allow() 221 GUEST_ASSERT(guest_exception_count == 0); in guest_code_filter_allow() 254 GUEST_ASSERT(data == 2); in guest_code_filter_allow() 265 GUEST_ASSERT(data == 0); in guest_code_filter_allow() 282 GUEST_ASSERT(data == 2); in guest_code_filter_allow() 334 GUEST_ASSERT(data == MSR_FS_BASE); in guest_code_permission_bitmap() 336 GUEST_ASSERT(data != MSR_GS_BASE); in guest_code_permission_bitmap() 342 GUEST_ASSERT(data != MSR_FS_BASE); in guest_code_permission_bitmap() 344 GUEST_ASSERT(data == MSR_GS_BASE); in guest_code_permission_bitmap() 348 GUEST_ASSERT(data != MSR_FS_BASE); in guest_code_permission_bitmap() [all …]
|
| A D | hyperv_svm_test.c | 82 GUEST_ASSERT(svm->vmcb_gpa); in guest_code() 97 GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL); in guest_code() 105 GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_MSR); in guest_code() 111 GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_MSR); in guest_code() 120 GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL); in guest_code() 126 GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_MSR); in guest_code() 135 GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_MSR); in guest_code() 140 GUEST_ASSERT(vmcb->control.exit_code == HV_SVM_EXITCODE_ENL); in guest_code() 141 GUEST_ASSERT(vmcb->control.exit_info_1 == HV_SVM_ENL_EXITCODE_TRAP_AFTER_FLUSH); in guest_code() 144 GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL); in guest_code()
|
| A D | vmx_apic_access_test.c | 43 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code() 44 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code() 59 GUEST_ASSERT(!vmlaunch()); in l1_guest_code() 60 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == EXIT_REASON_VMCALL); in l1_guest_code() 66 GUEST_ASSERT(!vmresume()); in l1_guest_code() 67 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == EXIT_REASON_VMCALL); in l1_guest_code()
|
| A D | vmx_invalid_nested_guest_state.c | 31 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code() 32 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code() 43 GUEST_ASSERT(!(vmreadz(CPU_BASED_VM_EXEC_CONTROL) & CPU_BASED_ACTIVATE_SECONDARY_CONTROLS) || in l1_guest_code() 46 GUEST_ASSERT(!vmlaunch()); in l1_guest_code() 49 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == EXIT_REASON_TRIPLE_FAULT); in l1_guest_code()
|
| A D | triple_fault_event_test.c | 30 GUEST_ASSERT(vmx->vmcs_gpa); in l1_guest_code_vmx() 31 GUEST_ASSERT(prepare_for_vmx_operation(vmx)); in l1_guest_code_vmx() 32 GUEST_ASSERT(load_vmcs(vmx)); in l1_guest_code_vmx() 37 GUEST_ASSERT(!vmlaunch()); in l1_guest_code_vmx() 39 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == EXIT_REASON_TRIPLE_FAULT); in l1_guest_code_vmx() 56 GUEST_ASSERT(0); in l1_guest_code_svm()
|
| A D | vmx_close_while_nested_test.c | 37 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code() 38 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code() 44 GUEST_ASSERT(!vmlaunch()); in l1_guest_code() 45 GUEST_ASSERT(0); in l1_guest_code()
|
| A D | cr4_cpuid_sync_test.c | 32 GUEST_ASSERT(get_cr4() & X86_CR4_OSXSAVE); in guest_code() 35 GUEST_ASSERT(this_cpu_has(X86_FEATURE_OSXSAVE)); in guest_code() 54 GUEST_ASSERT(!(regs[X86_FEATURE_OSXSAVE.reg] & BIT(X86_FEATURE_OSXSAVE.bit))); in guest_code() 57 GUEST_ASSERT(this_cpu_has(X86_FEATURE_OSXSAVE)); in guest_code()
|
| A D | fix_hypercall_test.c | 59 GUEST_ASSERT(0); in guest_main() 75 GUEST_ASSERT(ret == (uint64_t)-EFAULT); in guest_main() 76 GUEST_ASSERT(!memcmp(other_hypercall_insn, hypercall_insn, in guest_main() 79 GUEST_ASSERT(!ret); in guest_main() 80 GUEST_ASSERT(!memcmp(native_hypercall_insn, hypercall_insn, in guest_main()
|
| A D | vmx_dirty_log_test.c | 62 GUEST_ASSERT(vmx->vmcs_gpa); in l1_guest_code() 63 GUEST_ASSERT(prepare_for_vmx_operation(vmx)); in l1_guest_code() 64 GUEST_ASSERT(load_vmcs(vmx)); in l1_guest_code() 74 GUEST_ASSERT(!vmlaunch()); in l1_guest_code() 76 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == EXIT_REASON_VMCALL); in l1_guest_code()
|
| A D | sev_smoke_test.c | 23 GUEST_ASSERT(sev_msr & MSR_AMD64_SEV_ENABLED); in guest_snp_code() 24 GUEST_ASSERT(sev_msr & MSR_AMD64_SEV_ES_ENABLED); in guest_snp_code() 25 GUEST_ASSERT(sev_msr & MSR_AMD64_SEV_SNP_ENABLED); in guest_snp_code() 34 GUEST_ASSERT(rdmsr(MSR_AMD64_SEV) & MSR_AMD64_SEV_ENABLED); in guest_sev_es_code() 35 GUEST_ASSERT(rdmsr(MSR_AMD64_SEV) & MSR_AMD64_SEV_ES_ENABLED); in guest_sev_es_code() 47 GUEST_ASSERT(this_cpu_has(X86_FEATURE_SEV)); in guest_sev_code() 48 GUEST_ASSERT(rdmsr(MSR_AMD64_SEV) & MSR_AMD64_SEV_ENABLED); in guest_sev_code()
|
| A D | nested_emulation_test.c | 69 GUEST_ASSERT(prepare_for_vmx_operation(test_data)); in guest_code() 70 GUEST_ASSERT(load_vmcs(test_data)); in guest_code() 73 GUEST_ASSERT(!vmwrite(GUEST_IDTR_LIMIT, 0)); in guest_code() 74 GUEST_ASSERT(!vmwrite(GUEST_RIP, (u64)l2_guest_code)); in guest_code() 75 GUEST_ASSERT(!vmwrite(EXCEPTION_BITMAP, 0)); in guest_code()
|
| A D | svm_int_ctl_test.c | 47 GUEST_ASSERT(vintr_irq_called); in l2_guest_code() 48 GUEST_ASSERT(intr_irq_called); in l2_guest_code() 78 GUEST_ASSERT(vmcb->control.exit_code == SVM_EXIT_VMMCALL); in l1_guest_code()
|
| A D | vmx_nested_tsc_scaling_test.c | 90 GUEST_ASSERT(prepare_for_vmx_operation(vmx_pages)); in l1_guest_code() 91 GUEST_ASSERT(load_vmcs(vmx_pages)); in l1_guest_code() 110 GUEST_ASSERT(!vmlaunch()); in l1_guest_code() 111 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == EXIT_REASON_VMCALL); in l1_guest_code()
|
| A D | aperfmperf_test.c | 90 GUEST_ASSERT(!vmwrite(GUEST_RIP, (u64)l2_guest_code)); in l1_vmx_code() 91 GUEST_ASSERT(!vmlaunch()); in l1_vmx_code() 114 GUEST_ASSERT(vector == GP_VECTOR); in guest_no_aperfmperf() 117 GUEST_ASSERT(vector == GP_VECTOR); in guest_no_aperfmperf()
|
| /tools/testing/selftests/kvm/lib/arm64/ |
| A D | gic.c | 40 GUEST_ASSERT(gic_ops); in gic_dist_init() 56 GUEST_ASSERT(nr_cpus); in gic_init() 64 GUEST_ASSERT(gic_common_ops); in gic_irq_enable() 70 GUEST_ASSERT(gic_common_ops); in gic_irq_disable() 79 GUEST_ASSERT(gic_common_ops); in gic_get_and_ack_irq() 89 GUEST_ASSERT(gic_common_ops); in gic_set_eoi() 95 GUEST_ASSERT(gic_common_ops); in gic_set_dir() 101 GUEST_ASSERT(gic_common_ops); in gic_set_eoi_split() 107 GUEST_ASSERT(gic_common_ops); in gic_set_priority_mask() 113 GUEST_ASSERT(gic_common_ops); in gic_set_priority() [all …]
|
| A D | gic_v3.c | 48 GUEST_ASSERT(count--); in gicv3_gicd_wait_for_rwp() 64 GUEST_ASSERT(count--); in gicv3_gicr_wait_for_rwp() 89 GUEST_ASSERT(0); in get_intid_range() 177 GUEST_ASSERT(bits_per_field <= reg_bits); in gicv3_access_reg() 178 GUEST_ASSERT(!write || *val < (1U << bits_per_field)); in gicv3_access_reg() 183 GUEST_ASSERT(reg_bits == 32); in gicv3_access_reg() 228 GUEST_ASSERT(get_intid_range(intid) == SPI_RANGE); in gicv3_irq_set_config() 291 GUEST_ASSERT(count--); in gicv3_enable_redist() 302 GUEST_ASSERT(cpu < gicv3_data.nr_cpus); in gicv3_cpu_init() 369 GUEST_ASSERT(nr_cpus <= GICV3_MAX_CPUS); in gicv3_init()
|
| /tools/testing/selftests/kvm/lib/x86/ |
| A D | memstress.c | 38 GUEST_ASSERT(vmx->vmcs_gpa); in memstress_l1_guest_code() 39 GUEST_ASSERT(prepare_for_vmx_operation(vmx)); in memstress_l1_guest_code() 40 GUEST_ASSERT(load_vmcs(vmx)); in memstress_l1_guest_code() 41 GUEST_ASSERT(ept_1g_pages_supported()); in memstress_l1_guest_code() 47 GUEST_ASSERT(!vmlaunch()); in memstress_l1_guest_code() 48 GUEST_ASSERT(vmreadz(VM_EXIT_REASON) == EXIT_REASON_VMCALL); in memstress_l1_guest_code()
|
| /tools/testing/selftests/kvm/arm64/ |
| A D | psci_test.c | 129 GUEST_ASSERT(!psci_cpu_on(target_cpu, CPU_ON_ENTRY_ADDR, CPU_ON_CONTEXT_ID)); in guest_test_cpu_on() 134 GUEST_ASSERT((target_state == PSCI_0_2_AFFINITY_LEVEL_ON) || in guest_test_cpu_on() 171 GUEST_ASSERT(!psci_features(PSCI_1_0_FN_SYSTEM_SUSPEND)); in guest_test_system_suspend() 172 GUEST_ASSERT(!psci_features(PSCI_1_0_FN64_SYSTEM_SUSPEND)); in guest_test_system_suspend() 205 GUEST_ASSERT(psci_features(PSCI_1_3_FN_SYSTEM_OFF2) & in guest_test_system_off2() 207 GUEST_ASSERT(psci_features(PSCI_1_3_FN64_SYSTEM_OFF2) & in guest_test_system_off2() 212 GUEST_ASSERT(ret == PSCI_RET_INVALID_PARAMS); in guest_test_system_off2() 220 GUEST_ASSERT(ret == PSCI_RET_INTERNAL_FAILURE); in guest_test_system_off2() 228 GUEST_ASSERT(ret == PSCI_RET_INVALID_PARAMS); in guest_test_system_off2() 231 GUEST_ASSERT(ret == PSCI_RET_INTERNAL_FAILURE); in guest_test_system_off2()
|