/Linux-v5.15/include/asm-generic/ |
D | barrier.h | 68 #ifndef smp_rmb 69 #define smp_rmb() __smp_rmb() macro 82 #ifndef smp_rmb 83 #define smp_rmb() barrier() macro 200 #define smp_acquire__after_ctrl_dep() smp_rmb()
|
/Linux-v5.15/arch/x86/kvm/ |
D | irq.h | 73 smp_rmb(); in irqchip_split() 82 smp_rmb(); in irqchip_kernel() 96 smp_rmb(); in irqchip_in_kernel()
|
/Linux-v5.15/tools/include/asm/ |
D | barrier.h | 38 #ifndef smp_rmb 39 # define smp_rmb() rmb() macro
|
/Linux-v5.15/arch/arm64/include/asm/vdso/ |
D | compat_barrier.h | 35 #undef smp_rmb 39 #define smp_rmb() aarch32_smp_rmb() macro
|
/Linux-v5.15/tools/memory-model/litmus-tests/ |
D | MP+fencewmbonceonce+fencermbonceonce.litmus | 6 * This litmus test demonstrates that smp_wmb() and smp_rmb() provide 26 smp_rmb();
|
/Linux-v5.15/arch/sparc/include/asm/ |
D | vvar.h | 51 smp_rmb(); /* Finish all reads before we return seq */ in vvar_read_begin() 58 smp_rmb(); /* Finish all reads before checking the value of seq */ in vvar_read_retry()
|
/Linux-v5.15/include/vdso/ |
D | helpers.h | 16 smp_rmb(); in vdso_read_begin() 25 smp_rmb(); in vdso_read_retry()
|
/Linux-v5.15/include/clocksource/ |
D | hyperv_timer.h | 67 smp_rmb(); in hv_read_tsc_page_tsc() 77 smp_rmb(); in hv_read_tsc_page_tsc()
|
/Linux-v5.15/include/drm/ |
D | spsc_queue.h | 94 smp_rmb(); in spsc_queue_pop() 111 smp_rmb(); in spsc_queue_pop()
|
/Linux-v5.15/arch/mips/kernel/ |
D | rtlx.c | 125 smp_rmb(); in rtlx_open() 134 smp_rmb(); in rtlx_open() 254 smp_rmb(); in rtlx_read() 297 smp_rmb(); in rtlx_write()
|
/Linux-v5.15/Documentation/RCU/ |
D | rculist_nulls.rst | 47 but a version with an additional memory barrier (smp_rmb()) 55 pos && ({ next = pos->next; smp_rmb(); prefetch(next); 1; }) && 63 And note the traditional hlist_for_each_entry_rcu() misses this smp_rmb():: 132 Avoiding extra smp_rmb() 135 With hlist_nulls we can avoid extra smp_rmb() in lockless_lookup()
|
/Linux-v5.15/tools/testing/selftests/kvm/ |
D | rseq_test.c | 259 smp_rmb(); in main() 262 smp_rmb(); in main()
|
/Linux-v5.15/arch/powerpc/kvm/ |
D | irq.h | 18 smp_rmb(); in irqchip_in_kernel()
|
/Linux-v5.15/Documentation/translations/zh_CN/core-api/ |
D | local_ops.rst | 126 的CPU上分别使用显式的 ``smp_wmb()`` 和 ``smp_rmb()`` 内存屏障。如果你使 129 也应有一个 ``smp_rmb()`` 。
|
/Linux-v5.15/tools/include/linux/ |
D | ring_buffer.h | 63 smp_rmb(); in ring_buffer_read_head()
|
/Linux-v5.15/drivers/cpuidle/ |
D | coupled.c | 304 smp_rmb(); in cpuidle_coupled_get_state() 492 smp_rmb(); in cpuidle_enter_state_coupled() 573 smp_rmb(); in cpuidle_enter_state_coupled()
|
/Linux-v5.15/tools/arch/x86/include/asm/ |
D | barrier.h | 27 #define smp_rmb() barrier() macro
|
/Linux-v5.15/arch/arm/plat-versatile/ |
D | platsmp.c | 95 smp_rmb(); in versatile_boot_secondary()
|
/Linux-v5.15/include/linux/ |
D | cnt32_to_63.h | 97 smp_rmb(); \
|
/Linux-v5.15/kernel/livepatch/ |
D | patch.c | 83 smp_rmb(); in klp_ftrace_handler() 93 smp_rmb(); in klp_ftrace_handler()
|
/Linux-v5.15/arch/powerpc/include/asm/book3s/64/ |
D | hash-64k.h | 104 smp_rmb(); in __real_pte() 210 smp_rmb(); in get_hpte_slot_array()
|
/Linux-v5.15/lib/ |
D | is_single_threaded.c | 46 smp_rmb(); in current_is_single_threaded()
|
/Linux-v5.15/drivers/firmware/tegra/ |
D | ivc.c | 282 smp_rmb(); in tegra_ivc_read_get_next_frame() 432 smp_rmb(); in tegra_ivc_notified() 471 smp_rmb(); in tegra_ivc_notified() 511 smp_rmb(); in tegra_ivc_notified()
|
/Linux-v5.15/kernel/printk/ |
D | printk_ringbuffer.c | 470 smp_rmb(); /* LMM(desc_read:B) */ in desc_read() 522 smp_rmb(); /* LMM(desc_read:D) */ in desc_read() 728 smp_rmb(); /* LMM(data_push_tail:B) */ in data_push_tail() 857 smp_rmb(); /* LMM(desc_push_tail:C) */ in desc_push_tail() 909 smp_rmb(); /* LMM(desc_reserve:B) */ in desc_reserve() 1861 smp_rmb(); /* LMM(prb_first_seq:C) */ in prb_first_seq()
|
/Linux-v5.15/tools/memory-model/Documentation/ |
D | recipes.txt | 291 and to use smp_load_acquire() instead of smp_rmb(). However, the older 292 smp_wmb() and smp_rmb() APIs are still heavily used, so it is important 306 smp_rmb(); 311 smp_rmb() macro orders prior loads against later loads. Therefore, if 326 smp_rmb(); 335 * (A) smp_rmb() (C) 342 write side and smp_rmb() on the read side. 345 or smp_rmb(), any code fragment that would work with smp_rmb() and 390 * (A) smp_rmb() (C)
|