Lines Matching refs:MSR_VSX
128 newmsr |= MSR_VSX; in msr_check_and_set()
146 newmsr &= ~MSR_VSX; in __msr_check_and_clear()
162 msr &= ~MSR_VSX; in __giveup_fpu()
245 msr &= ~MSR_VSX; in __giveup_altivec()
311 WARN_ON((msr & MSR_VSX) && !((msr & MSR_FP) && (msr & MSR_VEC))); in __giveup_vsx()
324 msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx()
326 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx()
335 cpumsr = msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX); in enable_kernel_vsx()
338 (current->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP))) { in enable_kernel_vsx()
359 if (tsk->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP)) { in flush_vsx_to_thread()
416 msr_all_available |= MSR_VSX; in init_msr_all_available()
440 WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC))); in giveup_all()
536 if ((!(msr & MSR_VSX)) && should_restore_vsx()) { in restore_math()
538 new_msr |= MSR_VSX; in restore_math()
556 if (new_msr & MSR_VSX) in restore_math()
580 WARN_ON((usermsr & MSR_VSX) && !((usermsr & MSR_FP) && (usermsr & MSR_VEC))); in save_all()
1064 new->thread.regs->msr &= ~(MSR_FP | MSR_VEC | MSR_VSX); in tm_recheckpoint_new_task()
1118 msr_diff &= MSR_FP | MSR_VEC | MSR_VSX; in restore_tm_state()
1389 {MSR_VSX, "VSX"},
1709 childregs->msr &= ~(MSR_FP|MSR_VEC|MSR_VSX); in copy_thread()