Searched refs:VA_BITS (Results 1 – 21 of 21) sorted by relevance
50 #define VMEMMAP_SIZE (UL(1) << (VA_BITS - PAGE_SHIFT - 1 + STRUCT_PAGE_MAX_SHIFT))59 #define VA_BITS (CONFIG_ARM64_VA_BITS) macro61 (UL(1) << VA_BITS) + 1)63 (UL(1) << (VA_BITS - 1)) + 1)83 #define KASAN_SHADOW_SIZE (UL(1) << (VA_BITS - KASAN_SHADOW_SCALE_SHIFT))227 #define __is_lm_address(addr) (!!((addr) & BIT(VA_BITS - 1)))
75 return unlikely(idmap_t0sz != TCR_T0SZ(VA_BITS)); in __cpu_uses_extended_idmap()104 #define cpu_set_default_tcr_t0sz() __cpu_set_tcr_t0sz(TCR_T0SZ(VA_BITS))
82 return (image_addr & ~(SZ_1G - 1UL)) + (1UL << (VA_BITS - 1)); in efi_get_max_initrd_addr()
22 #define TASK_SIZE_64 (UL(1) << VA_BITS)
83 #define PTRS_PER_PGD (1 << (VA_BITS - PGDIR_SHIFT))
368 idmap_idx = hyp_idmap_start >> VA_BITS;
43 hyp_va_msb = idmap_addr & BIT(VA_BITS - 1); in compute_layout()44 hyp_va_msb ^= BIT(VA_BITS - 1); in compute_layout()49 if (kva_msb == (VA_BITS - 1)) { in compute_layout()55 va_mask = BIT(VA_BITS - 1) - 1; in compute_layout()69 tag_val = get_random_long() & GENMASK_ULL(VA_BITS - 2, tag_lsb); in compute_layout()
29 #define VA_BITS 44 macro31 #define VPTE_SIZE (1UL << (VA_BITS - PAGE_SHIFT + 3))33 #define VPTE_SIZE (1 << (VA_BITS - PAGE_SHIFT + 3))
553 #define VA_BITS 68 macro555 #define ESID_BITS (VA_BITS - (SID_SHIFT + CONTEXT_BITS))556 #define ESID_BITS_1T (VA_BITS - (SID_SHIFT_1T + CONTEXT_BITS))609 #define VSID_BITS_256M (VA_BITS - SID_SHIFT)617 #define VSID_BITS_1T (VA_BITS - SID_SHIFT_1T)710 unsigned long va_bits = VA_BITS; in get_vsid()
127 mask = ((1UL << (VA_BITS - 2)) - 1) & ~(SZ_2M - 1); in kaslr_early_init()128 offset = BIT(VA_BITS - 3) + (seed & mask); in kaslr_early_init()
12 VMCOREINFO_NUMBER(VA_BITS); in arch_crash_save_vmcoreinfo()
333 cmp x5, TCR_T0SZ(VA_BITS) // default T0SZ small enough?341 #if (VA_BITS < 48)352 #if VA_BITS != EXTRA_SHIFT
53 #define VA_BITS 32 macro
381 BUILD_BUG_ON(linear_region_size != BIT(VA_BITS - 1)); in arm64_memblock_init()
53 u64 idmap_t0sz = TCR_T0SZ(VA_BITS);672 if ((((long)addr) >> VA_BITS) != -1UL) in kern_addr_valid()
441 ldr x10, =TCR_TxSZ(VA_BITS) | TCR_CACHE_FLAGS | TCR_SMP_FLAGS | \
156 VA_BITS, mm->pgd); in show_pte()
974 unsigned long va_bits = VA_BITS; in vsid_unscramble()
793 if ((base ^ io_map_base) & BIT(VA_BITS - 1)) in __create_hyp_private_mapping()
1600 ias = min_t(unsigned long, ias, VA_BITS); in arm_smmu_domain_finalise()
196 # VA_BITS - PAGE_SHIFT - 3