Searched refs:shregs (Results 1 – 9 of 9) sorted by relevance
/Linux-v4.19/arch/powerpc/kvm/ |
D | book3s_hv_tm_builtin.c | 32 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation_early() 37 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early() 39 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation_early() 44 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early() 60 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation_early() 69 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early() 76 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation_early() 81 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation_early() 90 vcpu->arch.shregs.msr = (msr & ~MSR_TS_MASK) | MSR_TS_T; in kvmhv_p9_tm_emulation_early() 105 vcpu->arch.shregs.msr &= ~MSR_TS_MASK; /* go to N state */ in kvmhv_emulate_tm_rollback()
|
D | book3s_hv_tm.c | 20 u64 msr = vcpu->arch.shregs.msr; in emulate_tx_failure() 24 if (MSR_TM_SUSPENDED(vcpu->arch.shregs.msr)) in emulate_tx_failure() 46 u64 msr = vcpu->arch.shregs.msr; in kvmhv_p9_tm_emulation() 53 newmsr = vcpu->arch.shregs.srr1; in kvmhv_p9_tm_emulation() 59 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation() 61 vcpu->arch.regs.nip = vcpu->arch.shregs.srr0; in kvmhv_p9_tm_emulation() 92 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation() 108 vcpu->arch.shregs.msr = newmsr; in kvmhv_p9_tm_emulation() 143 vcpu->arch.shregs.msr = msr; in kvmhv_p9_tm_emulation() 179 vcpu->arch.shregs.msr &= ~MSR_TS_MASK; in kvmhv_p9_tm_emulation() [all …]
|
D | book3s_hv_ras.c | 74 unsigned long srr1 = vcpu->arch.shregs.msr; in kvmppc_realmode_mc_power7() 80 unsigned long dsisr = vcpu->arch.shregs.dsisr; in kvmppc_realmode_mc_power7()
|
D | book3s_hv.c | 332 vcpu->arch.shregs.msr = msr; in kvmppc_set_msr_hv() 400 vcpu->arch.regs.nip, vcpu->arch.shregs.msr, vcpu->arch.trap); in kvmppc_dump_regs() 408 vcpu->arch.shregs.srr0, vcpu->arch.shregs.srr1); in kvmppc_dump_regs() 410 vcpu->arch.shregs.sprg0, vcpu->arch.shregs.sprg1); in kvmppc_dump_regs() 412 vcpu->arch.shregs.sprg2, vcpu->arch.shregs.sprg3); in kvmppc_dump_regs() 414 vcpu->arch.cr, vcpu->arch.regs.xer, vcpu->arch.shregs.dsisr); in kvmppc_dump_regs() 415 pr_err("dar = %.16llx\n", vcpu->arch.shregs.dar); in kvmppc_dump_regs() 711 dt->srr1 = cpu_to_be64(vcpu->arch.shregs.msr); in kvmppc_create_dtl_entry() 1104 if (vcpu->arch.shregs.msr & MSR_HV) { in kvmppc_handle_exit_hv() 1108 vcpu->arch.shregs.msr); in kvmppc_handle_exit_hv() [all …]
|
D | book3s_hv_rm_mmu.c | 1095 key = (vcpu->arch.shregs.msr & MSR_PR) ? SLB_VSID_KP : SLB_VSID_KS; in kvmppc_hpte_hv_fault() 1112 if (data && (vcpu->arch.shregs.msr & MSR_DR)) { in kvmppc_hpte_hv_fault() 1148 if (data && (vcpu->arch.shregs.msr & MSR_IR)) in kvmppc_hpte_hv_fault()
|
D | book3s_64_mmu_hv.c | 295 if (MSR_TM_TRANSACTIONAL(vcpu->arch.shregs.msr)) in kvmppc_mmu_book3s_64_hv_reset_msr() 298 msr |= vcpu->arch.shregs.msr & MSR_TS_MASK; in kvmppc_mmu_book3s_64_hv_reset_msr() 362 int virtmode = vcpu->arch.shregs.msr & (data ? MSR_DR : MSR_IR); in kvmppc_mmu_book3s_64_hv_xlate() 400 key = (vcpu->arch.shregs.msr & MSR_PR) ? SLB_VSID_KP : SLB_VSID_KS; in kvmppc_mmu_book3s_64_hv_xlate()
|
D | trace_hv.h | 264 __entry->msr = vcpu->arch.shregs.msr;
|
/Linux-v4.19/arch/powerpc/kernel/ |
D | asm-offsets.c | 444 OFFSET(VCPU_MSR, kvm_vcpu, arch.shregs.msr); in main() 445 OFFSET(VCPU_SRR0, kvm_vcpu, arch.shregs.srr0); in main() 446 OFFSET(VCPU_SRR1, kvm_vcpu, arch.shregs.srr1); in main() 447 OFFSET(VCPU_SPRG0, kvm_vcpu, arch.shregs.sprg0); in main() 448 OFFSET(VCPU_SPRG1, kvm_vcpu, arch.shregs.sprg1); in main() 449 OFFSET(VCPU_SPRG2, kvm_vcpu, arch.shregs.sprg2); in main() 450 OFFSET(VCPU_SPRG3, kvm_vcpu, arch.shregs.sprg3); in main() 501 OFFSET(VCPU_DSISR, kvm_vcpu, arch.shregs.dsisr); in main() 502 OFFSET(VCPU_DAR, kvm_vcpu, arch.shregs.dar); in main()
|
/Linux-v4.19/arch/powerpc/include/asm/ |
D | kvm_host.h | 758 struct kvm_vcpu_arch_shared shregs; member
|