Searched refs:MSR_VSX (Results 1 – 22 of 22) sorted by relevance
| /linux/arch/powerpc/kernel/ |
| A D | process.c | 125 newmsr |= MSR_VSX; in msr_check_and_set() 143 newmsr &= ~MSR_VSX; in __msr_check_and_clear() 159 msr &= ~MSR_VSX; in __giveup_fpu() 242 msr &= ~MSR_VSX; in __giveup_altivec() 321 msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx() 323 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx() 413 msr_all_available |= MSR_VSX; in init_msr_all_available() 535 new_msr |= MSR_VSX; in restore_math() 553 if (new_msr & MSR_VSX) in restore_math() 1134 msr_diff &= MSR_FP | MSR_VEC | MSR_VSX; in restore_tm_state() [all …]
|
| A D | signal_64.c | 163 msr &= ~MSR_VSX; in __unsafe_setup_sigcontext() 176 msr |= MSR_VSX; in __unsafe_setup_sigcontext() 237 msr |= tsk->thread.ckpt_regs.msr & (MSR_FP | MSR_VEC | MSR_VSX); in setup_tm_sigcontexts() 302 if (msr & MSR_VSX) in setup_tm_sigcontexts() 310 msr |= MSR_VSX; in setup_tm_sigcontexts() 387 regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX)); in __unsafe_restore_sigcontext() 418 if ((msr & MSR_VSX) != 0) { in __unsafe_restore_sigcontext() 506 regs_set_return_msr(regs, regs->msr & ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX)); in restore_tm_sigcontexts() 552 if (v_regs && ((msr & MSR_VSX) != 0)) { in restore_tm_sigcontexts() 678 (new_msr & MSR_VSX)) in SYSCALL_DEFINE3()
|
| A D | signal_32.c | 302 msr &= ~MSR_VSX; in __unsafe_save_user_regs() 312 msr |= MSR_VSX; in __unsafe_save_user_regs() 431 if (msr & MSR_VSX) in save_tm_user_regs_unsafe() 436 msr |= MSR_VSX; in save_tm_user_regs_unsafe() 520 regs_set_return_msr(regs, regs->msr & ~MSR_VSX); in restore_user_regs() 521 if (msr & MSR_VSX) { in restore_user_regs() 621 regs_set_return_msr(regs, regs->msr & ~MSR_VSX); in restore_tm_user_regs() 622 if (msr & MSR_VSX) { in restore_tm_user_regs() 653 if (msr & MSR_VSX) { in restore_tm_user_regs() 1029 (new_msr & MSR_VSX)) in COMPAT_SYSCALL_DEFINE3()
|
| A D | fpu.S | 104 oris r5,r5,MSR_VSX@h
|
| A D | tm.S | 142 oris r15,r15, MSR_VSX@h 393 oris r5,r5, MSR_VSX@h
|
| A D | interrupt.c | 212 mathflags |= MSR_VEC | MSR_VSX; in interrupt_exit_user_prepare_main()
|
| A D | vector.S | 147 oris r12,r12,MSR_VSX@h
|
| A D | traps.c | 975 msr_mask = MSR_VSX; in p9_hmi_special_emu()
|
| A D | exceptions-64s.S | 2655 oris r10,r10,MSR_VSX@h
|
| /linux/arch/powerpc/lib/ |
| A D | ldstfp.S | 162 oris r7,r6,MSR_VSX@h 189 oris r7,r6,MSR_VSX@h
|
| A D | test_emulate_step.c | 72 regs->msr |= MSR_VSX; in init_pt_regs()
|
| A D | sstep.c | 3454 unsigned long msrbit = MSR_VSX; in emulate_loadstore() 3524 unsigned long msrbit = MSR_VSX; in emulate_loadstore()
|
| /linux/arch/powerpc/include/asm/ |
| A D | switch_to.h | 85 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in disable_kernel_vsx()
|
| A D | reg.h | 89 #define MSR_VSX __MASK(MSR_VSX_LG) /* Enable VSX */ macro
|
| /linux/arch/powerpc/kvm/ |
| A D | emulate_loadstore.c | 42 if (!(kvmppc_get_msr(vcpu) & MSR_VSX)) { in kvmppc_check_vsx_disabled() 317 MSR_VSX); in kvmppc_emulate_loadstore()
|
| A D | book3s_pr.c | 184 kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_core_vcpu_put_pr() 368 (MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_handle_lost_math_exts() 391 kvmppc_giveup_ext(vcpu, MSR_VSX); in kvmppc_save_tm_pr() 807 if (msr & MSR_VSX) in kvmppc_giveup_ext() 837 vcpu->arch.guest_owned_ext &= ~(msr | MSR_VSX); in kvmppc_giveup_ext() 875 if (msr == MSR_VSX) { in kvmppc_handle_ext() 889 msr = MSR_FP | MSR_VEC | MSR_VSX; in kvmppc_handle_ext() 1373 ext_msr = MSR_VSX; in kvmppc_handle_exit_pr() 1851 kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_vcpu_run_pr()
|
| A D | tm.S | 45 oris r8, r8, (MSR_VEC | MSR_VSX)@h 239 oris r5, r5, (MSR_VEC | MSR_VSX)@h
|
| A D | book3s_emulate.c | 177 kvmppc_giveup_ext(vcpu, MSR_VSX); in kvmppc_emulate_trchkpt()
|
| A D | book3s_hv_p9_entry.c | 506 msr_needed |= MSR_VSX; in kvmppc_msr_hard_disable_set_facilities()
|
| A D | book3s_hv_rmhandlers.S | 2454 oris r8,r8,MSR_VSX@h 2490 oris r8,r8,MSR_VSX@h
|
| A D | powerpc.c | 1205 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VSX); in kvmppc_complete_mmio_load()
|
| A D | book3s_hv.c | 5092 msr |= MSR_VSX; in kvmppc_vcpu_run_hv()
|
Completed in 63 milliseconds