| /arch/loongarch/kvm/ |
| A D | exit.c | 217 memcpy(run->iocsr_io.data, val, run->iocsr_io.len); in kvm_emu_iocsr() 281 struct kvm_run *run = vcpu->run; in kvm_trap_handle_gspr() local 373 struct kvm_run *run = vcpu->run; in kvm_emu_mmio_read() local 387 run->mmio.len = 4; in kvm_emu_mmio_read() 390 run->mmio.len = 8; in kvm_emu_mmio_read() 402 run->mmio.len = 1; in kvm_emu_mmio_read() 406 run->mmio.len = 1; in kvm_emu_mmio_read() 538 run->mmio.phys_addr, run->mmio.data); in kvm_complete_mmio_read() 548 struct kvm_run *run = vcpu->run; in kvm_emu_mmio_write() local 671 struct kvm_run *run = vcpu->run; in kvm_handle_rdwr_fault() local [all …]
|
| /arch/s390/kvm/ |
| A D | diag.c | 96 vcpu->run->s.regs.gprs[rx]); in __diag_page_ref_service() 98 if (vcpu->run->s.regs.gprs[rx] & 7) in __diag_page_ref_service() 117 vcpu->run->s.regs.gprs[ry] = 8; in __diag_page_ref_service() 131 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 145 vcpu->run->s.regs.gprs[ry] = 0; in __diag_page_ref_service() 151 vcpu->run->s.regs.gprs[ry] = 4; in __diag_page_ref_service() 244 vcpu->run->s390_reset_flags = 0; in __diag_ipl_functions() 261 vcpu->run->s390_reset_flags); in __diag_ipl_functions() 279 vcpu->run->s.regs.gprs[4]); in __diag_virtio_hypercall() 290 vcpu->run->s.regs.gprs[4]); in __diag_virtio_hypercall() [all …]
|
| A D | priv.c | 476 vcpu->run->s.regs.gprs[0] = 0; in handle_test_block() 554 if (vcpu->run->s.regs.gprs[1]) in handle_tsch() 556 vcpu->run->s.regs.gprs[1]); in handle_tsch() 645 reg0 = vcpu->run->s.regs.gprs[0]; in handle_pqap() 886 vcpu->run->s390_stsi.addr = addr; in insert_stsi_usr_data() 887 vcpu->run->s390_stsi.ar = ar; in insert_stsi_usr_data() 888 vcpu->run->s390_stsi.fc = fc; in insert_stsi_usr_data() 889 vcpu->run->s390_stsi.sel1 = sel1; in insert_stsi_usr_data() 890 vcpu->run->s390_stsi.sel2 = sel2; in insert_stsi_usr_data() 976 vcpu->run->s.regs.gprs[0] = 0; in handle_stsi() [all …]
|
| A D | kvm-s390.h | 25 static inline void kvm_s390_fpu_store(struct kvm_run *run) in kvm_s390_fpu_store() argument 27 fpu_stfpc(&run->s.regs.fpc); in kvm_s390_fpu_store() 29 save_vx_regs((__vector128 *)&run->s.regs.vrs); in kvm_s390_fpu_store() 31 save_fp_regs((freg_t *)&run->s.regs.fprs); in kvm_s390_fpu_store() 34 static inline void kvm_s390_fpu_load(struct kvm_run *run) in kvm_s390_fpu_load() argument 36 fpu_lfpc_safe(&run->s.regs.fpc); in kvm_s390_fpu_load() 38 load_vx_regs((__vector128 *)&run->s.regs.vrs); in kvm_s390_fpu_load() 40 load_fp_regs((freg_t *)&run->s.regs.fprs); in kvm_s390_fpu_load() 140 return (base2 ? vcpu->run->s.regs.gprs[base2] : 0) + disp2; in kvm_s390_get_base_disp_s() 155 return (base1 ? vcpu->run->s.regs.gprs[base1] : 0) + disp1; in kvm_s390_get_base_disp_siy() [all …]
|
| A D | sigp.c | 430 u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; in kvm_s390_handle_sigp() 443 parameter = vcpu->run->s.regs.gprs[r1]; in kvm_s390_handle_sigp() 445 parameter = vcpu->run->s.regs.gprs[r1 + 1]; in kvm_s390_handle_sigp() 452 &vcpu->run->s.regs.gprs[r1]); in kvm_s390_handle_sigp() 457 &vcpu->run->s.regs.gprs[r1]); in kvm_s390_handle_sigp() 479 u16 cpu_addr = vcpu->run->s.regs.gprs[r3]; in kvm_s390_handle_sigp_pei()
|
| /arch/mips/kvm/ |
| A D | emulate.c | 975 struct kvm_run *run = vcpu->run; in kvm_mips_emulate_store() local 999 run->mmio.len = 8; in kvm_mips_emulate_store() 1009 run->mmio.len = 4; in kvm_mips_emulate_store() 1018 run->mmio.len = 2; in kvm_mips_emulate_store() 1027 run->mmio.len = 1; in kvm_mips_emulate_store() 1038 run->mmio.len = 4; in kvm_mips_emulate_store() 1068 run->mmio.len = 4; in kvm_mips_emulate_store() 1253 run->mmio.phys_addr, run->mmio.len, data); in kvm_mips_emulate_store() 1271 struct kvm_run *run = vcpu->run; in kvm_mips_emulate_load() local 1498 run->mmio.phys_addr, run->mmio.len, run->mmio.data); in kvm_mips_emulate_load() [all …]
|
| /arch/riscv/kvm/ |
| A D | vcpu_insn.c | 265 run->riscv_csr.ret_value); in kvm_riscv_vcpu_csr_return() 317 run->riscv_csr.csr_num = csr_num; in csr_insn() 320 run->riscv_csr.ret_value = 0; in csr_insn() 337 run->riscv_csr.ret_value = val; in csr_insn() 571 run->mmio.is_write = false; in kvm_riscv_vcpu_mmio_load() 572 run->mmio.phys_addr = fault_addr; in kvm_riscv_vcpu_mmio_load() 573 run->mmio.len = len; in kvm_riscv_vcpu_mmio_load() 586 run->exit_reason = KVM_EXIT_MMIO; in kvm_riscv_vcpu_mmio_load() 704 run->mmio.is_write = true; in kvm_riscv_vcpu_mmio_store() 706 run->mmio.len = len; in kvm_riscv_vcpu_mmio_store() [all …]
|
| A D | vcpu_sbi.c | 128 run->riscv_sbi.args[0] = cp->a0; in kvm_riscv_vcpu_sbi_forward() 129 run->riscv_sbi.args[1] = cp->a1; in kvm_riscv_vcpu_sbi_forward() 130 run->riscv_sbi.args[2] = cp->a2; in kvm_riscv_vcpu_sbi_forward() 131 run->riscv_sbi.args[3] = cp->a3; in kvm_riscv_vcpu_sbi_forward() 135 run->riscv_sbi.ret[1] = 0; in kvm_riscv_vcpu_sbi_forward() 139 struct kvm_run *run, in kvm_riscv_vcpu_sbi_system_reset() argument 152 memset(&run->system_event, 0, sizeof(run->system_event)); in kvm_riscv_vcpu_sbi_system_reset() 153 run->system_event.type = type; in kvm_riscv_vcpu_sbi_system_reset() 154 run->system_event.ndata = 1; in kvm_riscv_vcpu_sbi_system_reset() 197 cp->a0 = run->riscv_sbi.ret[0]; in kvm_riscv_vcpu_sbi_return() [all …]
|
| A D | vcpu_exit.c | 15 static int gstage_page_fault(struct kvm_vcpu *vcpu, struct kvm_run *run, in gstage_page_fault() argument 34 return kvm_riscv_vcpu_mmio_load(vcpu, run, in gstage_page_fault() 38 return kvm_riscv_vcpu_mmio_store(vcpu, run, in gstage_page_fault() 187 int kvm_riscv_vcpu_exit(struct kvm_vcpu *vcpu, struct kvm_run *run, in kvm_riscv_vcpu_exit() argument 198 run->exit_reason = KVM_EXIT_UNKNOWN; in kvm_riscv_vcpu_exit() 230 ret = kvm_riscv_vcpu_virtual_insn(vcpu, run, trap); in kvm_riscv_vcpu_exit() 236 ret = gstage_page_fault(vcpu, run, trap); in kvm_riscv_vcpu_exit() 240 ret = kvm_riscv_vcpu_sbi_ecall(vcpu, run); in kvm_riscv_vcpu_exit() 243 run->exit_reason = KVM_EXIT_DEBUG; in kvm_riscv_vcpu_exit()
|
| A D | vcpu_sbi_replace.c | 17 static int kvm_sbi_ext_time_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, in kvm_sbi_ext_time_handler() argument 45 static int kvm_sbi_ext_ipi_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, in kvm_sbi_ext_ipi_handler() argument 92 static int kvm_sbi_ext_rfence_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, in kvm_sbi_ext_rfence_handler() argument 148 struct kvm_run *run, in kvm_sbi_ext_srst_handler() argument 160 kvm_riscv_vcpu_sbi_system_reset(vcpu, run, in kvm_sbi_ext_srst_handler() 167 kvm_riscv_vcpu_sbi_system_reset(vcpu, run, in kvm_sbi_ext_srst_handler() 190 struct kvm_run *run, in kvm_sbi_ext_dbcn_handler() argument 204 kvm_riscv_vcpu_sbi_forward(vcpu, run); in kvm_sbi_ext_dbcn_handler()
|
| A D | vcpu_sbi_base.c | 16 static int kvm_sbi_ext_base_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, in kvm_sbi_ext_base_handler() argument 44 kvm_riscv_vcpu_sbi_forward(vcpu, run); in kvm_sbi_ext_base_handler() 76 struct kvm_run *run, in kvm_sbi_ext_forward_handler() argument 83 kvm_riscv_vcpu_sbi_forward(vcpu, run); in kvm_sbi_ext_forward_handler()
|
| A D | vcpu_sbi_v01.c | 16 static int kvm_sbi_ext_v01_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, in kvm_sbi_ext_v01_handler() argument 35 kvm_riscv_vcpu_sbi_forward(vcpu, run); in kvm_sbi_ext_v01_handler() 65 kvm_riscv_vcpu_sbi_system_reset(vcpu, run, in kvm_sbi_ext_v01_handler()
|
| A D | vcpu_sbi_system.c | 12 static int kvm_sbi_ext_susp_handler(struct kvm_vcpu *vcpu, struct kvm_run *run, in kvm_sbi_ext_susp_handler() argument 50 kvm_riscv_vcpu_sbi_forward(vcpu, run); in kvm_sbi_ext_susp_handler()
|
| /arch/arm64/kvm/ |
| A D | mmio.c | 124 struct kvm_run *run = vcpu->run; in kvm_handle_mmio_return() local 127 data = kvm_mmio_read_buf(run->mmio.data, len); in kvm_handle_mmio_return() 155 struct kvm_run *run = vcpu->run; in io_mem_abort() local 179 run->exit_reason = KVM_EXIT_ARM_NISV; in io_mem_abort() 181 run->arm_nisv.fault_ipa = fault_ipa; in io_mem_abort() 215 run->mmio.is_write = is_write; in io_mem_abort() 216 run->mmio.phys_addr = fault_ipa; in io_mem_abort() 217 run->mmio.len = len; in io_mem_abort() 223 memcpy(run->mmio.data, data_buf, len); in io_mem_abort() 230 memcpy(run->mmio.data, data_buf, len); in io_mem_abort() [all …]
|
| A D | psci.c | 185 memset(&vcpu->run->system_event, 0, sizeof(vcpu->run->system_event)); in kvm_prepare_system_event() 186 vcpu->run->system_event.type = type; in kvm_prepare_system_event() 187 vcpu->run->system_event.ndata = 1; in kvm_prepare_system_event() 188 vcpu->run->system_event.data[0] = flags; in kvm_prepare_system_event() 189 vcpu->run->exit_reason = KVM_EXIT_SYSTEM_EVENT; in kvm_prepare_system_event() 216 struct kvm_run *run = vcpu->run; in kvm_psci_system_suspend() local 218 memset(&run->system_event, 0, sizeof(vcpu->run->system_event)); in kvm_psci_system_suspend() 219 run->system_event.type = KVM_SYSTEM_EVENT_SUSPEND; in kvm_psci_system_suspend() 220 run->exit_reason = KVM_EXIT_SYSTEM_EVENT; in kvm_psci_system_suspend()
|
| A D | handle_exit.c | 188 struct kvm_run *run = vcpu->run; in kvm_handle_guest_debug() local 194 run->exit_reason = KVM_EXIT_DEBUG; in kvm_handle_guest_debug() 195 run->debug.arch.hsr = lower_32_bits(esr); in kvm_handle_guest_debug() 196 run->debug.arch.hsr_high = upper_32_bits(esr); in kvm_handle_guest_debug() 197 run->flags = KVM_DEBUG_ARCH_HSR_HIGH_VALID; in kvm_handle_guest_debug() 201 run->debug.arch.far = vcpu->arch.fault.far_el2; in kvm_handle_guest_debug() 443 struct kvm_run *run = vcpu->run; in handle_exit() local 467 run->exit_reason = KVM_EXIT_FAIL_ENTRY; in handle_exit() 474 run->exit_reason = KVM_EXIT_FAIL_ENTRY; in handle_exit() 479 run->exit_reason = KVM_EXIT_INTERNAL_ERROR; in handle_exit()
|
| /arch/riscv/include/asm/ |
| A D | kvm_vcpu_insn.h | 36 int kvm_riscv_vcpu_csr_return(struct kvm_vcpu *vcpu, struct kvm_run *run); 37 int kvm_riscv_vcpu_virtual_insn(struct kvm_vcpu *vcpu, struct kvm_run *run, 40 int kvm_riscv_vcpu_mmio_load(struct kvm_vcpu *vcpu, struct kvm_run *run, 43 int kvm_riscv_vcpu_mmio_store(struct kvm_vcpu *vcpu, struct kvm_run *run, 46 int kvm_riscv_vcpu_mmio_return(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
| A D | kvm_vcpu_sbi.h | 47 int (*handler)(struct kvm_vcpu *vcpu, struct kvm_run *run, 64 void kvm_riscv_vcpu_sbi_forward(struct kvm_vcpu *vcpu, struct kvm_run *run); 66 struct kvm_run *run, 71 int kvm_riscv_vcpu_sbi_return(struct kvm_vcpu *vcpu, struct kvm_run *run); 83 int kvm_riscv_vcpu_sbi_ecall(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
| /arch/powerpc/kvm/ |
| A D | powerpc.c | 1135 struct kvm_run *run = vcpu->run; in kvmppc_complete_mmio_load() local 1142 switch (run->mmio.len) { in kvmppc_complete_mmio_load() 1247 struct kvm_run *run = vcpu->run; in __kvmppc_handle_load() local 1262 run->mmio.len = bytes; in __kvmppc_handle_load() 1263 run->mmio.is_write = 0; in __kvmppc_handle_load() 1333 struct kvm_run *run = vcpu->run; in kvmppc_handle_store() local 1349 run->mmio.len = bytes; in kvmppc_handle_store() 1350 run->mmio.is_write = 1; in kvmppc_handle_store() 1478 struct kvm_run *run = vcpu->run; in kvmppc_emulate_mmio_vsx_loadstore() local 1657 struct kvm_run *run = vcpu->run; in kvmppc_emulate_mmio_vmx_loadstore() local [all …]
|
| A D | booke.c | 751 vcpu->run->epr.epr = 0; in kvmppc_core_check_requests() 753 vcpu->run->exit_reason = KVM_EXIT_EPR; in kvmppc_core_check_requests() 866 struct kvm_run *run = vcpu->run; in kvmppc_handle_debug() local 897 run->debug.arch.status = 0; in kvmppc_handle_debug() 908 run->debug.arch.address = dbg_reg->dac1; in kvmppc_handle_debug() 1013 struct kvm_run *run = vcpu->run; in kvmppc_handle_exit() local 1071 run->exit_reason = KVM_EXIT_UNKNOWN; in kvmppc_handle_exit() 1072 run->ready_for_interrupt_injection = 1; in kvmppc_handle_exit() 1146 run->exit_reason = KVM_EXIT_DEBUG; in kvmppc_handle_exit() 1213 run->hw.hardware_exit_reason = exit_nr; in kvmppc_handle_exit() [all …]
|
| /arch/arm/mach-aspeed/ |
| A D | Kconfig | 10 Say Y here if you want to run your kernel on an ASpeed BMC SoC. 21 Say yes if you intend to run on an Aspeed ast2400 or similar 31 Say yes if you intend to run on an Aspeed ast2500 or similar 42 Say yes if you intend to run on an Aspeed ast2600 or similar
|
| /arch/x86/kvm/vmx/ |
| A D | tdx.c | 1122 vcpu->run->ex.error_code = 0; in tdx_handle_exception_nmi() 1156 if (vcpu->run->hypercall.ret) { in tdx_complete_vmcall_map_gpa() 1197 tdx->vcpu.run->hypercall.ret = 0; in __tdx_map_gpa() 1262 vcpu->run->system_event.ndata = 16; in tdx_report_fatal_error() 1438 vcpu->run->mmio.phys_addr = gpa; in tdx_emulate_mmio() 1439 vcpu->run->mmio.len = size; in tdx_emulate_mmio() 1440 vcpu->run->mmio.is_write = write; in tdx_emulate_mmio() 1491 vcpu->run->tdx.flags = 0; in tdx_get_td_vm_call_info() 1525 vcpu->run->tdx.flags = 0; in tdx_get_quote() 1547 vcpu->run->tdx.flags = 0; in tdx_setup_event_notify_interrupt() [all …]
|
| /arch/loongarch/include/asm/ |
| A D | kvm_vcpu.h | 44 int kvm_complete_mmio_read(struct kvm_vcpu *vcpu, struct kvm_run *run); 45 int kvm_complete_iocsr_read(struct kvm_vcpu *vcpu, struct kvm_run *run); 46 int kvm_complete_user_service(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
| /arch/x86/include/uapi/asm/ |
| A D | sgx.h | 158 struct sgx_enclave_run *run); 230 struct sgx_enclave_run *run);
|
| /arch/arm64/boot/dts/freescale/ |
| A D | fsl-lx2162a-sr-som.dtsi | 5 // Copyright 2021 Rabeeh Khoury <rabeeh@solid-run.com> 6 // Copyright 2023 Josua Mayer <josua@solid-run.com>
|