/Linux-v4.19/arch/x86/include/asm/ |
D | smap.h | 31 ALTERNATIVE "", __stringify(__ASM_CLAC), X86_FEATURE_SMAP 34 ALTERNATIVE "", __stringify(__ASM_STAC), X86_FEATURE_SMAP 63 ALTERNATIVE("", __stringify(__ASM_CLAC), X86_FEATURE_SMAP) 65 ALTERNATIVE("", __stringify(__ASM_STAC), X86_FEATURE_SMAP)
|
D | barrier.h | 15 #define mb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "mfence", \ 17 #define rmb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "lfence", \ 19 #define wmb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "sfence", \
|
D | alternative.h | 136 #define ALTERNATIVE(oldinstr, newinstr, feature) \ macro 169 asm volatile (ALTERNATIVE(oldinstr, newinstr, feature) : : : "memory") 186 asm volatile (ALTERNATIVE(oldinstr, newinstr, feature) \ 205 asm volatile (ALTERNATIVE(oldinstr, newinstr, feature) \ 210 asm volatile (ALTERNATIVE("call %P[old]", "call %P[new]", feature) \
|
D | nospec-branch.h | 144 ALTERNATIVE "jmp .Lskip_rsb_\@", \ 173 ALTERNATIVE( \ 187 ALTERNATIVE( \ 243 ALTERNATIVE("jmp 910f", 255 asm volatile(ALTERNATIVE("", "wrmsr", %c[feature])
|
D | arch_hweight.h | 27 asm (ALTERNATIVE("call __sw_hweight32", POPCNT32, X86_FEATURE_POPCNT) in __arch_hweight32() 55 asm (ALTERNATIVE("call __sw_hweight64", POPCNT64, X86_FEATURE_POPCNT) in __arch_hweight64()
|
D | kvm_para.h | 21 ALTERNATIVE(".byte 0x0f,0x01,0xc1", ".byte 0x0f,0x01,0xd9", X86_FEATURE_VMMCALL)
|
D | alternative-asm.h | 43 .macro ALTERNATIVE oldinstr, newinstr, feature
|
/Linux-v4.19/arch/x86/entry/ |
D | calling.h | 207 ALTERNATIVE "", "SET_NOFLUSH_BIT \reg", X86_FEATURE_PCID 213 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 224 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 227 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID 264 ALTERNATIVE "jmp .Ldone_\@", "", X86_FEATURE_PTI 282 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 284 ALTERNATIVE "jmp .Lwrcr3_\@", "", X86_FEATURE_PCID
|
D | entry_32.S | 161 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 171 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 192 ALTERNATIVE "jmp .Lend_\@", "", X86_FEATURE_PTI 301 ALTERNATIVE "jmp .Lswitched_\@", "", X86_FEATURE_PTI 320 ALTERNATIVE "jmp .Lend_\@", "", X86_BUG_ESPFIX 381 ALTERNATIVE "", "jmp .Lend_\@", X86_FEATURE_XENPV 538 ALTERNATIVE "", "jmp .Lend_\@", X86_FEATURE_XENPV 886 ALTERNATIVE "testl %eax, %eax; jz .Lsyscall_32_done", \ 1146 ALTERNATIVE "pushl $do_general_protection", \
|
D | entry_64_compat.S | 141 ALTERNATIVE "testl %eax, %eax; jz .Lsyscall_32_done", \ 259 ALTERNATIVE "testl %eax, %eax; jz .Lsyscall_32_done", \
|
/Linux-v4.19/arch/arm64/include/asm/ |
D | uaccess.h | 60 asm(ALTERNATIVE("nop", SET_PSTATE_UAO(1), ARM64_HAS_UAO)); in set_fs() 62 asm(ALTERNATIVE("nop", SET_PSTATE_UAO(0), ARM64_HAS_UAO, in set_fs() 190 asm(ALTERNATIVE("nop", SET_PSTATE_PAN(0), ARM64_HAS_PAN, in __uaccess_disable_hw_pan() 196 asm(ALTERNATIVE("nop", SET_PSTATE_PAN(1), ARM64_HAS_PAN, in __uaccess_enable_hw_pan() 203 asm(ALTERNATIVE("nop", SET_PSTATE_PAN(1), alt, \ 210 asm(ALTERNATIVE("nop", SET_PSTATE_PAN(0), alt, \ 267 "1:"ALTERNATIVE(instr " " reg "1, [%2]\n", \ 338 "1:"ALTERNATIVE(instr " " reg "1, [%2]\n", \
|
D | kvm_hyp.h | 30 asm volatile(ALTERNATIVE("mrs %0, " __stringify(r##nvh),\ 40 asm volatile(ALTERNATIVE("msr " __stringify(r##nvh) ", %x0",\ 54 asm volatile(ALTERNATIVE("mrs %0, " __stringify(r##_EL2),\ 64 asm volatile(ALTERNATIVE("msr " __stringify(r##_EL2) ", %x0",\ 116 asm volatile(ALTERNATIVE("nop \n", \
|
D | io.h | 67 asm volatile(ALTERNATIVE("ldrb %w0, [%1]", in __raw_readb() 79 asm volatile(ALTERNATIVE("ldrh %w0, [%1]", in __raw_readw() 90 asm volatile(ALTERNATIVE("ldr %w0, [%1]", in __raw_readl() 101 asm volatile(ALTERNATIVE("ldr %0, [%1]", in __raw_readq()
|
D | tlbflush.h | 41 ALTERNATIVE("nop\n nop", \ 48 ALTERNATIVE("nop\n nop", \
|
D | lse.h | 36 ALTERNATIVE(llsc, lse, ARM64_HAS_LSE_ATOMICS)
|
D | percpu.h | 27 asm volatile(ALTERNATIVE("msr tpidr_el1, %0", in set_my_cpu_offset() 41 asm(ALTERNATIVE("mrs %0, tpidr_el1", in __my_cpu_offset()
|
/Linux-v4.19/arch/s390/include/asm/ |
D | alternative.h | 109 #define ALTERNATIVE(oldinstr, altinstr, facility) \ macro 142 asm volatile(ALTERNATIVE(oldinstr, altinstr, facility) : : : "memory")
|
D | nospec-insn.h | 111 # The ALTERNATIVE replacement code has a .+10 which targets 118 ALTERNATIVE "ex %r0,_LC_BR_R1", ".insn ril,0xc60000000000,0,.+10", 35
|
D | alternative-asm.h | 67 .macro ALTERNATIVE oldinstr, newinstr, feature
|
D | spinlock.h | 93 ALTERNATIVE("", ".long 0xb2fa0070", 49) /* NIAI 7 */ in arch_spin_unlock()
|
/Linux-v4.19/arch/x86/entry/vdso/vdso32/ |
D | system_call.S | 60 ALTERNATIVE "", SYSENTER_SEQUENCE, X86_FEATURE_SEP
|
/Linux-v4.19/arch/s390/lib/ |
D | spinlock.c | 78 ALTERNATIVE("", ".long 0xb2fa0040", 49) /* NIAI 4 */ in arch_load_niai4() 89 ALTERNATIVE("", ".long 0xb2fa0080", 49) /* NIAI 8 */ in arch_cmpxchg_niai8()
|
/Linux-v4.19/arch/arm64/kernel/ |
D | process.c | 394 asm(ALTERNATIVE("nop", SET_PSTATE_UAO(1), ARM64_HAS_UAO)); in uao_thread_switch() 396 asm(ALTERNATIVE("nop", SET_PSTATE_UAO(0), ARM64_HAS_UAO)); in uao_thread_switch()
|
/Linux-v4.19/arch/x86/lib/ |
D | copy_page_64.S | 17 ALTERNATIVE "jmp copy_page_regs", "", X86_FEATURE_REP_GOOD
|
/Linux-v4.19/arch/arm64/mm/ |
D | context.c | 251 asm(ALTERNATIVE("nop; nop; nop", in post_ttbr_update_workaround()
|