Searched refs:MSR_SPE (Results 1 – 10 of 10) sorted by relevance
374 msr_check_and_set(MSR_SPE); in giveup_spe()376 msr_check_and_clear(MSR_SPE); in giveup_spe()384 msr_check_and_set(MSR_SPE); in enable_kernel_spe()386 if (current->thread.regs && (current->thread.regs->msr & MSR_SPE)) { in enable_kernel_spe()397 if (tsk->thread.regs->msr & MSR_SPE) { in flush_spe_to_thread()418 msr_all_available |= MSR_SPE; in init_msr_all_available()446 if (usermsr & MSR_SPE) in giveup_all()588 if (usermsr & MSR_SPE) in save_all()601 if (tsk->thread.regs->msr & MSR_SPE) in flush_all_to_thread()
306 msr |= MSR_SPE; in save_user_regs()454 msr |= MSR_SPE; in save_tm_user_regs()561 regs->msr &= ~MSR_SPE; in restore_user_regs()562 if (msr & MSR_SPE) { in restore_user_regs()672 regs->msr &= ~MSR_SPE; in restore_tm_user_regs()673 if (msr & MSR_SPE) { in restore_tm_user_regs()
868 oris r5,r5,MSR_SPE@h872 oris r9,r9,MSR_SPE@h888 oris r3,r3,MSR_SPE@h1012 lis r3,MSR_SPE@h
726 oris r0,r0,MSR_SPE@h /* Disable SPE */
2170 if (regs->msr & MSR_SPE) in SPEFloatingPointRoundException()
83 msr_check_and_clear(MSR_SPE); in disable_kernel_spe()
87 #ifndef MSR_SPE88 #define MSR_SPE 0 macro
30 #define MSR_SPE __MASK(MSR_SPE_LG) macro
93 vcpu->arch.shadow_msr &= ~MSR_SPE; in kvmppc_vcpu_disable_spe()103 vcpu->arch.shadow_msr |= MSR_SPE; in kvmppc_vcpu_enable_spe()109 if (vcpu->arch.shared->msr & MSR_SPE) { in kvmppc_vcpu_sync_spe()110 if (!(vcpu->arch.shadow_msr & MSR_SPE)) in kvmppc_vcpu_sync_spe()112 } else if (vcpu->arch.shadow_msr & MSR_SPE) { in kvmppc_vcpu_sync_spe()1129 if (vcpu->arch.shared->msr & MSR_SPE) in kvmppc_handle_exit()
310 if (vcpu->arch.shadow_msr & MSR_SPE) in kvmppc_core_vcpu_put_e500()