Lines Matching refs:MSR_VEC

242 	msr &= ~MSR_VEC;  in __giveup_altivec()
252 msr_check_and_set(MSR_VEC); in giveup_altivec()
254 msr_check_and_clear(MSR_VEC); in giveup_altivec()
264 cpumsr = msr_check_and_set(MSR_VEC); in enable_kernel_altivec()
266 if (current->thread.regs && (current->thread.regs->msr & MSR_VEC)) { in enable_kernel_altivec()
291 if (tsk->thread.regs->msr & MSR_VEC) { in flush_altivec_to_thread()
310 WARN_ON((msr & MSR_VSX) && !((msr & MSR_FP) && (msr & MSR_VEC))); in __giveup_vsx()
315 if (msr & MSR_VEC) in __giveup_vsx()
323 msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx()
325 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx()
334 cpumsr = msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX); in enable_kernel_vsx()
337 (current->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP))) { in enable_kernel_vsx()
358 if (tsk->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP)) { in flush_vsx_to_thread()
413 msr_all_available |= MSR_VEC; in init_msr_all_available()
439 WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC))); in giveup_all()
443 if (usermsr & MSR_VEC) in giveup_all()
532 if ((!(msr & MSR_VEC)) && should_restore_altivec()) in restore_math()
533 new_msr |= MSR_VEC; in restore_math()
536 if (((msr | new_msr) & (MSR_FP | MSR_VEC)) == (MSR_FP | MSR_VEC)) in restore_math()
552 if (new_msr & MSR_VEC) in restore_math()
579 WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC))); in save_all()
584 if (usermsr & MSR_VEC) in save_all()
972 if ((thr->ckpt_regs.msr & MSR_VEC) == 0) in tm_reclaim_thread()
1082 new->thread.regs->msr &= ~(MSR_FP | MSR_VEC | MSR_VSX); in tm_recheckpoint_new_task()
1136 msr_diff &= MSR_FP | MSR_VEC | MSR_VSX; in restore_tm_state()
1142 if (cpu_has_feature(CPU_FTR_ALTIVEC) && msr_diff & MSR_VEC) in restore_tm_state()
1201 if (usermsr & MSR_VEC) in kvmppc_save_user_regs()
1466 {MSR_VEC, "VEC"},
1802 childregs->msr &= ~(MSR_FP|MSR_VEC|MSR_VSX); in copy_thread()