Searched refs:SID_SHIFT (Results 1 – 14 of 14) sorted by relevance
| /Linux-v5.4/arch/powerpc/include/asm/book3s/64/ |
| D | mmu-hash.h | 263 return SID_SHIFT; in segment_shift() 441 mask = (1ul << (SID_SHIFT - VPN_SHIFT)) - 1; in hpt_hash() 442 hash = (vpn >> (SID_SHIFT - VPN_SHIFT)) ^ in hpt_hash() 568 #define ESID_BITS (VA_BITS - (SID_SHIFT + CONTEXT_BITS)) 608 #define MAX_USER_CONTEXT_65BIT_VA ((ASM_CONST(1) << (65 - (SID_SHIFT + ESID_BITS))) - 2) 636 #define VSID_BITS_256M (VA_BITS - SID_SHIFT) 637 #define VSID_BITS_65_256M (65 - SID_SHIFT) 650 #define USER_VSID_RANGE (1UL << (ESID_BITS + SID_SHIFT)) 780 vsid_bits = va_bits - SID_SHIFT; in get_vsid() 782 ((ea >> SID_SHIFT) & ESID_BITS_MASK); in get_vsid()
|
| /Linux-v5.4/arch/powerpc/kvm/ |
| D | book3s_32_mmu.c | 36 #ifndef SID_SHIFT 37 #define SID_SHIFT 28 macro 89 kvmppc_mmu_book3s_32_esid_to_vsid(vcpu, eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_book3s_32_ea_to_vp() 163 eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_book3s_32_xlate_bat() 340 kvmppc_mmu_map_segment(vcpu, srnum << SID_SHIFT); in kvmppc_mmu_book3s_32_mtsrin() 356 ulong ea = esid << SID_SHIFT; in kvmppc_mmu_book3s_32_esid_to_vsid()
|
| D | book3s_64_mmu.c | 79 return slbe->tb ? SID_SHIFT_1T : SID_SHIFT; in kvmppc_slb_sid_shift() 424 kvmppc_mmu_map_segment(vcpu, esid << SID_SHIFT); in kvmppc_mmu_book3s_64_slbmte() 583 (mp_ea >> SID_SHIFT) == esid; in segment_contains_magic_page() 590 ulong ea = esid << SID_SHIFT; in kvmppc_mmu_book3s_64_esid_to_vsid() 603 gvsid <<= SID_SHIFT_1T - SID_SHIFT; in kvmppc_mmu_book3s_64_esid_to_vsid() 604 gvsid |= esid & ((1ul << (SID_SHIFT_1T - SID_SHIFT)) - 1); in kvmppc_mmu_book3s_64_esid_to_vsid() 652 unlikely(esid == (mp_ea >> SID_SHIFT)) && in kvmppc_mmu_book3s_64_esid_to_vsid()
|
| D | book3s_32_mmu_host.c | 158 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page() 167 vpn = (vsid << (SID_SHIFT - VPN_SHIFT)) | in kvmppc_mmu_map_page() 304 u32 esid = eaddr >> SID_SHIFT; in kvmppc_mmu_map_segment()
|
| D | book3s_64_mmu_host.c | 106 vcpu->arch.mmu.esid_to_vsid(vcpu, orig_pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_map_page() 217 vcpu->arch.mmu.esid_to_vsid(vcpu, pte->eaddr >> SID_SHIFT, &vsid); in kvmppc_mmu_unmap_page() 313 u64 esid = eaddr >> SID_SHIFT; in kvmppc_mmu_map_segment()
|
| D | book3s_pr.c | 700 pte.vpage |= ((u64)VSID_REAL << (SID_SHIFT - 12)); in kvmppc_handle_pagefault() 709 vcpu->arch.mmu.esid_to_vsid(vcpu, eaddr >> SID_SHIFT, &vsid); in kvmppc_handle_pagefault() 712 pte.vpage |= ((u64)VSID_REAL_DR << (SID_SHIFT - 12)); in kvmppc_handle_pagefault() 714 pte.vpage |= ((u64)VSID_REAL_IR << (SID_SHIFT - 12)); in kvmppc_handle_pagefault() 1152 sr = svcpu->sr[kvmppc_get_pc(vcpu) >> SID_SHIFT]; in kvmppc_handle_exit_pr() 1198 sr = svcpu->sr[dar >> SID_SHIFT]; in kvmppc_handle_exit_pr()
|
| /Linux-v5.4/arch/powerpc/include/asm/ |
| D | page_64.h | 27 #define SID_SHIFT 28 macro 30 #define GET_ESID(x) (((x) >> SID_SHIFT) & SID_MASK)
|
| D | kvm_book3s_32.h | 31 #define SID_SHIFT 28 macro
|
| D | kvm_book3s_64.h | 338 va_low ^= v >> (SID_SHIFT - 16); in compute_tlbie_rb()
|
| /Linux-v5.4/arch/powerpc/mm/book3s64/ |
| D | slb.c | 292 esid = ea >> SID_SHIFT; in preload_add() 437 get_paca()->slb_cache[i] << SID_SHIFT; in switch_slb() 496 ea = (unsigned long)ti->slb_preload_esid[idx] << SID_SHIFT; in switch_slb()
|
| D | hash_native.c | 728 *vpn = vsid << (SID_SHIFT - VPN_SHIFT) | seg_off >> VPN_SHIFT; in hpte_decode()
|
| /Linux-v5.4/drivers/misc/cxl/ |
| D | fault.c | 43 hash = (slb->esid >> SID_SHIFT) & mask; in find_free_sste()
|
| /Linux-v5.4/arch/powerpc/kernel/ |
| D | setup_64.c | 637 return 1UL << SID_SHIFT; in ppc64_bolted_size()
|
| /Linux-v5.4/arch/powerpc/platforms/pseries/ |
| D | lpar.c | 1924 vsid_modulus = ((1UL << (va_bits - SID_SHIFT)) - 1); in vsid_unscramble()
|