Searched refs:shregs (Results 1 – 13 of 13) sorted by relevance
| /linux/arch/powerpc/kvm/ |
| A D | book3s_hv_tm_builtin.c | 40 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation_early() 45 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early() 47 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation_early() 52 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early() 68 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation_early() 77 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early() 84 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early() 90 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early() 99 vcpu->arch.shregs.msr = (msr & ~MSR_TS_MASK) | MSR_TS_T; in kvmhv_p9_tm_emulation_early() 115 vcpu->arch.shregs.msr &= ~MSR_TS_MASK; /* go to N state */ in kvmhv_emulate_tm_rollback()
|
| A D | book3s_hv_tm.c | 19 u64 msr = vcpu->arch.shregs.msr; in emulate_tx_failure() 23 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) in emulate_tx_failure() 45 u64 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation() 72 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation() 78 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation() 80 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation() 112 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation() 128 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation() 166 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation() 205 vcpu->arch.shregs.msr &= ~MSR_TS_MASK; in kvmhv_p9_tm_emulation() [all …]
|
| A D | book3s_hv_p9_entry.c | 664 mtspr(SPRN_SPRG0, vcpu->arch.shregs.sprg0); in kvmhv_vcpu_entry_p9() 665 mtspr(SPRN_SPRG1, vcpu->arch.shregs.sprg1); in kvmhv_vcpu_entry_p9() 666 mtspr(SPRN_SPRG2, vcpu->arch.shregs.sprg2); in kvmhv_vcpu_entry_p9() 716 mtspr(SPRN_DAR, vcpu->arch.shregs.dar); in kvmhv_vcpu_entry_p9() 718 mtspr(SPRN_SRR0, vcpu->arch.shregs.srr0); in kvmhv_vcpu_entry_p9() 719 mtspr(SPRN_SRR1, vcpu->arch.shregs.srr1); in kvmhv_vcpu_entry_p9() 730 vcpu->arch.shregs.srr0 = mfspr(SPRN_SRR0); in kvmhv_vcpu_entry_p9() 731 vcpu->arch.shregs.srr1 = mfspr(SPRN_SRR1); in kvmhv_vcpu_entry_p9() 732 vcpu->arch.shregs.dar = mfspr(SPRN_DAR); in kvmhv_vcpu_entry_p9() 808 (vcpu->arch.shregs.msr & MSR_TS_S)) { in kvmhv_vcpu_entry_p9() [all …]
|
| A D | book3s_hv_nested.c | 45 hr->srr0 = vcpu->arch.shregs.srr0; in kvmhv_save_hv_regs() 46 hr->srr1 = vcpu->arch.shregs.srr1; in kvmhv_save_hv_regs() 47 hr->sprg[0] = vcpu->arch.shregs.sprg0; in kvmhv_save_hv_regs() 48 hr->sprg[1] = vcpu->arch.shregs.sprg1; in kvmhv_save_hv_regs() 49 hr->sprg[2] = vcpu->arch.shregs.sprg2; in kvmhv_save_hv_regs() 113 hr->srr0 = vcpu->arch.shregs.srr0; in save_hv_return_state() 114 hr->srr1 = vcpu->arch.shregs.srr1; in save_hv_return_state() 155 vcpu->arch.shregs.srr0 = hr->srr0; in restore_hv_regs() 156 vcpu->arch.shregs.srr1 = hr->srr1; in restore_hv_regs() 183 vcpu->arch.shregs.srr0 = hr->srr0; in kvmhv_restore_hv_return_state() [all …]
|
| A D | book3s_hv_nestedv2.c | 285 vcpu->arch.shregs.srr0); in gs_msg_ops_vcpu_fill_info() 289 vcpu->arch.shregs.srr1); in gs_msg_ops_vcpu_fill_info() 293 vcpu->arch.shregs.sprg0); in gs_msg_ops_vcpu_fill_info() 297 vcpu->arch.shregs.sprg1); in gs_msg_ops_vcpu_fill_info() 301 vcpu->arch.shregs.sprg2); in gs_msg_ops_vcpu_fill_info() 305 vcpu->arch.shregs.sprg3); in gs_msg_ops_vcpu_fill_info() 309 vcpu->arch.shregs.dar); in gs_msg_ops_vcpu_fill_info() 313 vcpu->arch.shregs.dsisr); in gs_msg_ops_vcpu_fill_info() 317 vcpu->arch.shregs.msr); in gs_msg_ops_vcpu_fill_info() 557 vcpu->arch.shregs.dar = kvmppc_gse_get_u64(gse); in gs_msg_ops_vcpu_refresh_info() [all …]
|
| A D | book3s_hv.h | 58 vcpu->arch.shregs.msr = val; in __kvmppc_set_msr_hv() 65 return vcpu->arch.shregs.msr; in __kvmppc_get_msr_hv()
|
| A D | book3s_hv_ras.c | 71 unsigned long srr1 = vcpu->arch.shregs.msr; in kvmppc_realmode_mc_power7() 76 unsigned long dsisr = vcpu->arch.shregs.dsisr; in kvmppc_realmode_mc_power7()
|
| A D | book3s_hv.c | 497 vcpu->arch.shregs.srr0, vcpu->arch.shregs.srr1); in kvmppc_dump_regs() 499 vcpu->arch.shregs.sprg0, vcpu->arch.shregs.sprg1); in kvmppc_dump_regs() 501 vcpu->arch.shregs.sprg2, vcpu->arch.shregs.sprg3); in kvmppc_dump_regs() 810 dt->srr1 = cpu_to_be64(vcpu->arch.shregs.msr); in __kvmppc_create_dtl_entry() 1647 vcpu->arch.shregs.msr); in kvmppc_handle_exit_hv() 2996 vcpu->arch.shared = &vcpu->arch.shregs; in kvmppc_core_vcpu_create_hv() 4301 vcpu->arch.regs.msr = vcpu->arch.shregs.msr; in kvmhv_vcpu_entry_p9_nested() 4328 mtspr(SPRN_DAR, vcpu->arch.shregs.dar); in kvmhv_vcpu_entry_p9_nested() 4329 mtspr(SPRN_DSISR, vcpu->arch.shregs.dsisr); in kvmhv_vcpu_entry_p9_nested() 4337 vcpu->arch.shregs.msr = vcpu->arch.regs.msr; in kvmhv_vcpu_entry_p9_nested() [all …]
|
| A D | book3s_hv_rm_mmu.c | 1241 key = (vcpu->arch.shregs.msr & MSR_PR) ? SLB_VSID_KP : SLB_VSID_KS; in kvmppc_hpte_hv_fault() 1258 if (data && (vcpu->arch.shregs.msr & MSR_DR)) { in kvmppc_hpte_hv_fault() 1294 if (data && (vcpu->arch.shregs.msr & MSR_IR)) in kvmppc_hpte_hv_fault()
|
| A D | book3s_hv_builtin.c | 584 if (vcpu->arch.shregs.msr & MSR_EE) { in kvmppc_guest_entry_inject_int()
|
| A D | trace_hv.h | 282 __entry->msr = vcpu->arch.shregs.msr;
|
| /linux/arch/powerpc/kernel/ |
| A D | asm-offsets.c | 376 OFFSET(VCPU_MSR, kvm_vcpu, arch.shregs.msr); in main() 377 OFFSET(VCPU_SRR0, kvm_vcpu, arch.shregs.srr0); in main() 378 OFFSET(VCPU_SRR1, kvm_vcpu, arch.shregs.srr1); in main() 379 OFFSET(VCPU_SPRG0, kvm_vcpu, arch.shregs.sprg0); in main() 380 OFFSET(VCPU_SPRG1, kvm_vcpu, arch.shregs.sprg1); in main() 381 OFFSET(VCPU_SPRG2, kvm_vcpu, arch.shregs.sprg2); in main() 382 OFFSET(VCPU_SPRG3, kvm_vcpu, arch.shregs.sprg3); in main() 430 OFFSET(VCPU_DSISR, kvm_vcpu, arch.shregs.dsisr); in main() 431 OFFSET(VCPU_DAR, kvm_vcpu, arch.shregs.dar); in main()
|
| /linux/arch/powerpc/include/asm/ |
| A D | kvm_host.h | 818 struct kvm_vcpu_arch_shared shregs; member
|
Completed in 38 milliseconds