Home
last modified time | relevance | path

Searched refs:MSR_PR (Results 1 – 25 of 41) sorted by relevance

12

/Linux-v4.19/arch/powerpc/kvm/
Dbook3s_hv_tm_builtin.c45 if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) in kvmhv_p9_tm_emulation_early()
49 ((msr & MSR_PR) && !(mfspr(SPRN_FSCR) & FSCR_EBB))) in kvmhv_p9_tm_emulation_early()
83 if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) in kvmhv_p9_tm_emulation_early()
Dbook3s_hv_tm.c26 if (msr & MSR_PR) { in emulate_tx_failure()
65 if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) { in kvmhv_p9_tm_emulation()
76 if ((msr & MSR_PR) && !(vcpu->arch.fscr & FSCR_EBB)) { in kvmhv_p9_tm_emulation()
113 if ((msr & MSR_PR) && (vcpu->arch.vcore->pcr & PCR_ARCH_206)) { in kvmhv_p9_tm_emulation()
Dbook3s_emulate.c92 if ((kvmppc_get_msr(vcpu) & MSR_PR) && level > PRIV_PROBLEM) in spr_allowed()
160 if (kvmppc_get_msr(vcpu) & MSR_PR) in kvmppc_emulate_treclaim()
235 if (guest_msr & MSR_PR) in kvmppc_emulate_tabort()
376 if ((kvmppc_get_msr(vcpu) & MSR_PR) || in kvmppc_core_emulate_op_pr()
495 if (!(kvmppc_get_msr(vcpu) & MSR_PR)) { in kvmppc_core_emulate_op_pr()
537 WARN_ON(guest_msr & MSR_PR); in kvmppc_core_emulate_op_pr()
560 if (guest_msr & MSR_PR) { in kvmppc_core_emulate_op_pr()
594 if (guest_msr & MSR_PR) { in kvmppc_core_emulate_op_pr()
843 if (kvmppc_get_msr(vcpu) & MSR_PR) { in kvmppc_core_emulate_mtspr_pr()
848 if ((kvmppc_get_msr(vcpu) & MSR_PR) || sprn == 0) { in kvmppc_core_emulate_mtspr_pr()
[all …]
Dbook3s_32_mmu.c157 if (kvmppc_get_msr(vcpu) & MSR_PR) { in kvmppc_mmu_book3s_32_xlate_bat()
240 if ((sr_kp(sre) && (kvmppc_get_msr(vcpu) & MSR_PR)) || in kvmppc_mmu_book3s_32_xlate_pte()
241 (sr_ks(sre) && !(kvmppc_get_msr(vcpu) & MSR_PR))) in kvmppc_mmu_book3s_32_xlate_pte()
318 !(kvmppc_get_msr(vcpu) & MSR_PR)) { in kvmppc_mmu_book3s_32_xlate()
400 if (msr & MSR_PR) in kvmppc_mmu_book3s_32_esid_to_vsid()
Dbook3s_64_mmu.c237 !(kvmppc_get_msr(vcpu) & MSR_PR)) { in kvmppc_mmu_book3s_64_xlate()
282 if ((kvmppc_get_msr(vcpu) & MSR_PR) && slbe->Kp) in kvmppc_mmu_book3s_64_xlate()
284 else if (!(kvmppc_get_msr(vcpu) & MSR_PR) && slbe->Ks) in kvmppc_mmu_book3s_64_xlate()
327 !(kvmppc_get_msr(vcpu) & MSR_PR)) in kvmppc_mmu_book3s_64_xlate()
580 return mp_ea && !(kvmppc_get_msr(vcpu) & MSR_PR) && in segment_contains_magic_page()
641 if (kvmppc_get_msr(vcpu) & MSR_PR) in kvmppc_mmu_book3s_64_esid_to_vsid()
651 !(kvmppc_get_msr(vcpu) & MSR_PR)) { in kvmppc_mmu_book3s_64_esid_to_vsid()
Dbook3s_pr.c204 smsr |= MSR_ME | MSR_RI | MSR_IR | MSR_DR | MSR_PR | MSR_EE; in kvmppc_recalc_shadow_msr()
217 if (!(guest_msr & MSR_PR)) in kvmppc_recalc_shadow_msr()
286 if (unlikely((old_msr & MSR_PR) && in kvmppc_copy_from_svcpu()
471 if (!(msr & MSR_PR) && MSR_TM_TRANSACTIONAL(msr)) in kvmppc_set_msr_pr()
498 if ((kvmppc_get_msr(vcpu) & (MSR_PR|MSR_IR|MSR_DR)) != in kvmppc_set_msr_pr()
499 (old_msr & (MSR_PR|MSR_IR|MSR_DR))) { in kvmppc_set_msr_pr()
504 if (!(msr & MSR_PR) && vcpu->arch.magic_page_pa) { in kvmppc_set_msr_pr()
523 !(old_msr & MSR_PR) && !(old_msr & MSR_SF) && (msr & MSR_SF)) { in kvmppc_set_msr_pr()
960 if (!(kvmppc_get_msr(vcpu) & MSR_PR)) in kvmppc_emulate_fac()
1018 if ((fac == FSCR_TM_LG) && !(kvmppc_get_msr(vcpu) & MSR_PR)) in kvmppc_handle_fac()
[all …]
Dbook3s_32_mmu_host.c96 if (kvmppc_get_msr(vcpu) & MSR_PR) in find_sid_vsid()
282 if (kvmppc_get_msr(vcpu) & MSR_PR) in create_sid_map()
Dbook3s_64_mmu_host.c62 if (kvmppc_get_msr(vcpu) & MSR_PR) in find_sid_vsid()
243 if (kvmppc_get_msr(vcpu) & MSR_PR) in create_sid_map()
De500.h223 return !!(vcpu->arch.shared->msr & MSR_PR); in get_cur_pr()
Dbooke.c420 crit = crit && !(vcpu->arch.shared->msr & MSR_PR); in kvmppc_booke_irqprio_deliver()
1114 if (vcpu->arch.shared->msr & (MSR_PR | MSR_GS)) { in kvmppc_handle_exit()
1218 if (!(vcpu->arch.shared->msr & MSR_PR)) { in kvmppc_handle_exit()
1232 if (!(vcpu->arch.shared->msr & MSR_PR) && in kvmppc_handle_exit()
1253 if (!(vcpu->arch.shared->msr & MSR_PR) && in kvmppc_handle_exit()
1980 if (!(vcpu->arch.shared->msr & MSR_PR) && in kvmppc_xlate()
De500_mmu_host.c314 u32 pr = vcpu->arch.shared->msr & MSR_PR; in kvmppc_e500_setup_stlbe()
675 pr = vcpu->arch.shared->msr & MSR_PR; in kvmppc_load_last_inst()
Dbook3s_64_mmu_radix.c118 if (kvmppc_get_msr(vcpu) & MSR_PR) { in kvmppc_mmu_radix_xlate()
/Linux-v4.19/arch/powerpc/include/asm/
Dreg_booke.h46 #define MSR_USER32 (MSR_ | MSR_PR | MSR_EE)
50 #define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE)
53 #define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE)
Dptrace.h108 #define user_mode(regs) (((regs)->msr & MSR_PR) != 0)
Dreg.h97 #define MSR_PR __MASK(MSR_PR_LG) /* Problem State / Privilege Level */ macro
139 #define MSR_USER32 (MSR_ | MSR_PR | MSR_EE)
144 #define MSR_USER (MSR_KERNEL|MSR_PR|MSR_EE)
Dexception-64s.h495 andi. r10,r12,MSR_PR; /* See if coming from user */ \
/Linux-v4.19/arch/powerpc/kernel/
Dexceptions-64e.S77 andi. r3,r3,MSR_PR
154 andi. r3,r3,MSR_PR
296 andi. r10,r11,MSR_PR; /* save stack pointer */ \
627 andi. r0,r12,MSR_PR;
646 andi. r0,r12,MSR_PR;
771 1: andi. r14,r11,MSR_PR; /* check for userspace again */
841 1: andi. r14,r11,MSR_PR; /* check for userspace again */
1078 andi. r6,r10,MSR_PR
Dentry_64.S72 andi. r10,r12,MSR_PR
279 andi. r6,r8,MSR_PR
738 andi. r3,r3,MSR_PR
962 andi. r0,r3,MSR_PR
Dhead_booke.h43 andi. r11, r11, MSR_PR; /* check whether user or kernel */\
131 andi. r11,r11,MSR_PR; \
Dhead_40x.S115 andi. r11,r11,MSR_PR; \
154 andi. r11,r11,MSR_PR; \
698 andi. r10,r9,MSR_IR|MSR_PR /* check supervisor + MMU off */
Dexceptions-64s.S466 andi. r11,r12,MSR_PR /* See if coming from user. */
655 andi. r9,r11,MSR_PR // Check for exception from userspace
656 cmpdi cr4,r9,MSR_PR // And save the result in CR4 for later
826 li r10,MSR_PR /* Build a mask of MSR_PR .. */
834 andi. r10,r12,MSR_PR; /* Set CR0 correctly for label */
Dentry_32.S153 andi. r2,r9,MSR_PR
439 andi. r4,r8,MSR_PR
817 andi. r0,r3,MSR_PR
1068 andi. r3,r3,MSR_PR; \
/Linux-v4.19/arch/powerpc/lib/
Dsstep.c1236 if (regs->msr & MSR_PR) in analyse_instr()
1490 if (regs->msr & MSR_PR) in analyse_instr()
1496 if (regs->msr & MSR_PR) in analyse_instr()
1504 if (regs->msr & MSR_PR) in analyse_instr()
2857 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_FP)) in emulate_loadstore()
2864 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_VEC)) in emulate_loadstore()
2879 if (!(regs->msr & MSR_PR) && !(regs->msr & msrbit)) in emulate_loadstore()
2916 !(regs->msr & MSR_PR) && in emulate_loadstore()
2928 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_FP)) in emulate_loadstore()
2935 if (!(regs->msr & MSR_PR) && !(regs->msr & MSR_VEC)) in emulate_loadstore()
[all …]
/Linux-v4.19/arch/powerpc/xmon/
Dxmon.c511 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) in xmon_core()
656 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) { in xmon_core()
706 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) in xmon_bpt()
737 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) in xmon_break_match()
747 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) != (MSR_IR|MSR_64BIT)) in xmon_iabr_match()
772 if ((regs->msr & (MSR_IR|MSR_PR|MSR_64BIT)) == (MSR_IR|MSR_64BIT)) { in xmon_fault_handler()
1111 if ((regs->msr & (MSR_64BIT|MSR_PR|MSR_IR)) == (MSR_64BIT|MSR_IR)) { in do_step()
1599 if (regs->msr & MSR_PR) in print_bug_trap()
/Linux-v4.19/arch/powerpc/platforms/pseries/
Dras.c409 (MSR_LE|MSR_RI|MSR_DR|MSR_IR|MSR_ME|MSR_PR| in pSeries_system_reset_exception()

12