Searched refs:MSR_VSX (Results 1 – 19 of 19) sorted by relevance
/Linux-v4.19/arch/powerpc/kernel/ |
D | process.c | 149 newmsr |= MSR_VSX; in msr_check_and_set() 168 newmsr &= ~MSR_VSX; in __msr_check_and_clear() 186 msr &= ~MSR_VSX; in __giveup_fpu() 282 msr &= ~MSR_VSX; in __giveup_altivec() 364 WARN_ON((msr & MSR_VSX) && !((msr & MSR_FP) && (msr & MSR_VEC))); in __giveup_vsx() 377 msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx() 379 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx() 388 cpumsr = msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX); in enable_kernel_vsx() 391 (current->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP))) { in enable_kernel_vsx() 411 if (tsk->thread.regs->msr & (MSR_VSX|MSR_VEC|MSR_FP)) { in flush_vsx_to_thread() [all …]
|
D | signal_64.c | 154 msr &= ~MSR_VSX; in setup_sigcontext() 168 msr |= MSR_VSX; in setup_sigcontext() 227 msr |= tsk->thread.ckpt_regs.msr & (MSR_FP | MSR_VEC | MSR_VSX); in setup_tm_sigcontexts() 299 if (msr & MSR_VSX) in setup_tm_sigcontexts() 307 msr |= MSR_VSX; in setup_tm_sigcontexts() 380 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX); in restore_sigcontext() 413 if ((msr & MSR_VSX) != 0) { in restore_sigcontext() 512 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX); in restore_tm_sigcontexts() 559 if (v_regs && ((msr & MSR_VSX) != 0)) { in restore_tm_sigcontexts() 650 (new_msr & MSR_VSX)) in SYSCALL_DEFINE3()
|
D | signal_32.c | 431 msr &= ~MSR_VSX; in save_user_regs() 443 msr |= MSR_VSX; in save_user_regs() 588 if (msr & MSR_VSX) { in save_tm_user_regs() 597 msr |= MSR_VSX; in save_tm_user_regs() 696 regs->msr &= ~MSR_VSX; in restore_user_regs() 697 if (msr & MSR_VSX) { in restore_user_regs() 808 regs->msr &= ~MSR_VSX; in restore_tm_user_regs() 809 if (msr & MSR_VSX) { in restore_tm_user_regs() 1081 (new_msr & MSR_VSX)) in COMPAT_SYSCALL_DEFINE3()
|
D | fpu.S | 92 oris r5,r5,MSR_VSX@h
|
D | tm.S | 134 oris r15,r15, MSR_VSX@h 369 oris r5,r5, MSR_VSX@h
|
D | vector.S | 131 oris r12,r12,MSR_VSX@h
|
D | traps.c | 877 msr_mask = MSR_VSX; in p9_hmi_special_emu()
|
D | exceptions-64s.S | 1361 oris r10,r10,MSR_VSX@h
|
/Linux-v4.19/arch/powerpc/lib/ |
D | ldstfp.S | 168 oris r7,r6,MSR_VSX@h 195 oris r7,r6,MSR_VSX@h
|
D | test_emulate_step.c | 69 regs->msr |= MSR_VSX; in init_pt_regs()
|
D | sstep.c | 2871 unsigned long msrbit = MSR_VSX; in emulate_loadstore() 2942 unsigned long msrbit = MSR_VSX; in emulate_loadstore()
|
/Linux-v4.19/arch/powerpc/include/asm/ |
D | switch_to.h | 61 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in disable_kernel_vsx()
|
D | reg.h | 90 #define MSR_VSX __MASK(MSR_VSX_LG) /* Enable VSX */ macro
|
/Linux-v4.19/arch/powerpc/kvm/ |
D | emulate_loadstore.c | 53 if (!(kvmppc_get_msr(vcpu) & MSR_VSX)) { in kvmppc_check_vsx_disabled() 340 MSR_VSX); in kvmppc_emulate_loadstore()
|
D | book3s_pr.c | 139 kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_core_vcpu_put_pr() 328 (MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_handle_lost_math_exts() 351 kvmppc_giveup_ext(vcpu, MSR_VSX); in kvmppc_save_tm_pr() 794 if (msr & MSR_VSX) in kvmppc_giveup_ext() 824 vcpu->arch.guest_owned_ext &= ~(msr | MSR_VSX); in kvmppc_giveup_ext() 862 if (msr == MSR_VSX) { in kvmppc_handle_ext() 876 msr = MSR_FP | MSR_VEC | MSR_VSX; in kvmppc_handle_ext() 1358 ext_msr = MSR_VSX; in kvmppc_handle_exit_pr() 1850 kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_vcpu_run_pr()
|
D | tm.S | 45 oris r8, r8, (MSR_VEC | MSR_VSX)@h 231 oris r5, r5, (MSR_VEC | MSR_VSX)@h
|
D | book3s_emulate.c | 192 kvmppc_giveup_ext(vcpu, MSR_VSX); in kvmppc_emulate_trchkpt()
|
D | book3s_hv_rmhandlers.S | 3075 oris r8,r8,MSR_VSX@h 3110 oris r8,r8,MSR_VSX@h
|
D | powerpc.c | 1159 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VSX); in kvmppc_complete_mmio_load()
|