Lines Matching +full:0 +full:x23
42 #if (PAGE_OFFSET & 0x1fffff) != 0
67 .quad 0 // Image load offset from start of RAM, little-endian
70 .quad 0 // reserved
71 .quad 0 // reserved
72 .quad 0 // reserved
86 * x23 primary_entry() .. start_kernel() physical misalignment/KASLR offset
94 adrp x23, __PHYS_OFFSET
95 and x23, x23, MIN_KIMG_ALIGN - 1 // KASLR offset, defaults to 0
121 add x1, x0, #0x20 // 4 x 8 bytes
231 mov \count, #0
298 and x6, x6, #(0xf << ID_AA64MMFR2_LVA_SHIFT)
367 add x5, x5, x23 // add KASLR displacement
457 tst x23, ~(MIN_KIMG_ALIGN - 1) // already running randomized?
458 b.ne 0f
460 cbz x0, 0f // KASLR disabled? just proceed
461 orr x23, x23, x0 // record KASLR offset
464 0:
557 hvc #0
600 .quad 0
708 update_early_cpu_boot_status 0, x2, x3
729 and x0, x0, #(0xf << ID_AA64MMFR2_LVA_SHIFT)
762 add x11, x11, x23 // actual virtual offset
766 0: cmp x9, x10
771 b.ne 0b
772 add x14, x14, x23 // relocate
773 str x14, [x12, x23]
774 b 0b
819 sub x15, x23, x24 // delta from previous offset
821 mov x24, x23 // save new offset
826 tbnz x11, #0, 3f // branch to handle bitmaps
827 add x13, x11, x23
836 tbz x11, #0, 5f // skip bit if not set
868 mov x24, #0 // no RELR displacement yet
877 * If we return here, we have a KASLR displacement in x23 which we need