Searched refs:MSR_VEC (Results 1 – 19 of 19) sorted by relevance
| /Linux-v4.19/arch/powerpc/kernel/ |
| D | process.c | 119 (tsk->thread.ckpt_regs.msr & MSR_VEC); in tm_active_with_altivec() 279 msr &= ~MSR_VEC; in __giveup_altivec() 291 msr_check_and_set(MSR_VEC); in giveup_altivec() 293 msr_check_and_clear(MSR_VEC); in giveup_altivec() 303 cpumsr = msr_check_and_set(MSR_VEC); in enable_kernel_altivec() 305 if (current->thread.regs && (current->thread.regs->msr & MSR_VEC)) { in enable_kernel_altivec() 329 if (tsk->thread.regs->msr & MSR_VEC) { in flush_altivec_to_thread() 364 WARN_ON((msr & MSR_VSX) && !((msr & MSR_FP) && (msr & MSR_VEC))); in __giveup_vsx() 369 if (msr & MSR_VEC) in __giveup_vsx() 377 msr_check_and_set(MSR_FP|MSR_VEC|MSR_VSX); in giveup_vsx() [all …]
|
| D | signal_64.c | 131 msr |= MSR_VEC; in setup_sigcontext() 227 msr |= tsk->thread.ckpt_regs.msr & (MSR_FP | MSR_VEC | MSR_VSX); in setup_tm_sigcontexts() 248 if (msr & MSR_VEC) in setup_tm_sigcontexts() 260 msr |= MSR_VEC; in setup_tm_sigcontexts() 268 if (msr & MSR_VEC) in setup_tm_sigcontexts() 380 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX); in restore_sigcontext() 389 if (v_regs != NULL && (msr & MSR_VEC) != 0) { in restore_sigcontext() 512 regs->msr &= ~(MSR_FP | MSR_FE0 | MSR_FE1 | MSR_VEC | MSR_VSX); in restore_tm_sigcontexts() 525 if (v_regs != NULL && tm_v_regs != NULL && (msr & MSR_VEC) != 0) { in restore_tm_sigcontexts() 578 msr_check_and_set(msr & (MSR_FP | MSR_VEC)); in restore_tm_sigcontexts() [all …]
|
| D | vector.S | 49 oris r5,r5,MSR_VEC@h 68 oris r9,r9,MSR_VEC@h 72 oris r12,r12,MSR_VEC@h 123 andis. r5,r12,MSR_VEC@h
|
| D | signal_32.c | 409 msr |= MSR_VEC; in save_user_regs() 529 if (msr & MSR_VEC) { in save_tm_user_regs() 544 msr |= MSR_VEC; in save_tm_user_regs() 557 if (msr & MSR_VEC) { in save_tm_user_regs() 671 regs->msr &= ~MSR_VEC; in restore_user_regs() 672 if (msr & MSR_VEC) { in restore_user_regs() 774 regs->msr &= ~MSR_VEC; in restore_tm_user_regs() 775 if (msr & MSR_VEC) { in restore_tm_user_regs() 864 msr_check_and_set(msr & (MSR_FP | MSR_VEC)); in restore_tm_user_regs() 870 if (msr & MSR_VEC) { in restore_tm_user_regs() [all …]
|
| D | tm.S | 131 oris r15, r15, MSR_VEC@h 365 oris r5, r5, MSR_VEC@h
|
| D | entry_64.S | 254 andis. r0,r8,MSR_VEC@h
|
| D | entry_32.S | 674 oris r0,r0,MSR_VEC@h /* Disable altivec */
|
| D | traps.c | 879 msr_mask = MSR_VEC; in p9_hmi_special_emu()
|
| /Linux-v4.19/arch/powerpc/include/asm/ |
| D | switch_to.h | 49 msr_check_and_clear(MSR_VEC); in disable_kernel_altivec() 61 msr_check_and_clear(MSR_FP|MSR_VEC|MSR_VSX); in disable_kernel_vsx()
|
| D | reg.h | 89 #define MSR_VEC __MASK(MSR_VEC_LG) /* Enable AltiVec */ macro
|
| /Linux-v4.19/arch/powerpc/lib/ |
| D | ldstfp.S | 78 oris r7, r6, MSR_VEC@h 102 oris r7, r6, MSR_VEC@h
|
| D | sstep.c | 580 if (regs->msr & MSR_VEC) in do_vec_load() 603 if (regs->msr & MSR_VEC) in do_vec_store() 843 if (regs->msr & MSR_VEC) in do_vsx_load() 874 if (regs->msr & MSR_VEC) in do_vsx_store() 2864 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_VEC)) in emulate_loadstore() 2878 msrbit = MSR_VEC; in emulate_loadstore() 2935 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_VEC)) in emulate_loadstore() 2949 msrbit = MSR_VEC; in emulate_loadstore()
|
| D | test_emulate_step.c | 68 regs->msr |= MSR_VEC; in init_pt_regs()
|
| /Linux-v4.19/arch/powerpc/kvm/ |
| D | book3s_pr.c | 139 kvmppc_giveup_ext(vcpu, MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_core_vcpu_put_pr() 328 (MSR_FP | MSR_VEC | MSR_VSX); in kvmppc_handle_lost_math_exts() 335 else if (ext_diff == MSR_VEC) in kvmppc_handle_lost_math_exts() 795 msr |= MSR_FP | MSR_VEC; in kvmppc_giveup_ext() 817 if (msr & MSR_VEC) { in kvmppc_giveup_ext() 818 if (current->thread.regs->msr & MSR_VEC) in kvmppc_giveup_ext() 876 msr = MSR_FP | MSR_VEC | MSR_VSX; in kvmppc_handle_ext() 897 if (msr & MSR_VEC) { in kvmppc_handle_ext() 935 if (lost_ext & MSR_VEC) { in kvmppc_handle_lost_ext() 1354 ext_msr = MSR_VEC; in kvmppc_handle_exit_pr() [all …]
|
| D | emulate_loadstore.c | 65 if (!(kvmppc_get_msr(vcpu) & MSR_VEC)) { in kvmppc_check_altivec_disabled() 295 MSR_VEC); in kvmppc_emulate_loadstore()
|
| D | tm.S | 45 oris r8, r8, (MSR_VEC | MSR_VSX)@h 231 oris r5, r5, (MSR_VEC | MSR_VSX)@h
|
| D | booke.c | 188 if (!(current->thread.regs->msr & MSR_VEC)) { in kvmppc_load_guest_altivec() 193 current->thread.regs->msr |= MSR_VEC; in kvmppc_load_guest_altivec() 207 if (current->thread.regs->msr & MSR_VEC) in kvmppc_save_guest_altivec()
|
| D | book3s_hv_rmhandlers.S | 3070 oris r8,r8,MSR_VEC@h 3105 oris r8,r8,MSR_VEC@h
|
| D | powerpc.c | 1176 vcpu->kvm->arch.kvm_ops->giveup_ext(vcpu, MSR_VEC); in kvmppc_complete_mmio_load()
|