/Linux-v4.19/include/asm-generic/ |
D | barrier.h | 79 #ifndef smp_rmb 80 #define smp_rmb() __smp_rmb() macro 97 #ifndef smp_rmb 98 #define smp_rmb() barrier() macro 220 #define smp_acquire__after_ctrl_dep() smp_rmb()
|
/Linux-v4.19/arch/x86/kvm/ |
D | irq.h | 86 smp_rmb(); in pic_in_kernel() 95 smp_rmb(); in irqchip_split() 104 smp_rmb(); in irqchip_kernel() 113 smp_rmb(); in irqchip_in_kernel()
|
/Linux-v4.19/tools/memory-model/litmus-tests/ |
D | MP+fencewmbonceonce+fencermbonceonce.litmus | 6 * This litmus test demonstrates that smp_wmb() and smp_rmb() provide 26 smp_rmb();
|
/Linux-v4.19/arch/sparc/include/asm/ |
D | vvar.h | 50 smp_rmb(); /* Finish all reads before we return seq */ in vvar_read_begin() 57 smp_rmb(); /* Finish all reads before checking the value of seq */ in vvar_read_retry()
|
/Linux-v4.19/arch/mips/include/asm/ |
D | vdso.h | 105 smp_rmb(); in vdso_data_read_begin() 117 smp_rmb(); in vdso_data_read_retry()
|
/Linux-v4.19/arch/x86/include/asm/ |
D | vgtod.h | 57 smp_rmb(); in gtod_read_begin() 64 smp_rmb(); in gtod_read_retry()
|
D | pgtable-3level.h | 84 smp_rmb(); in pmd_read_atomic() 327 smp_rmb(); in gup_get_pte() 329 smp_rmb(); in gup_get_pte()
|
/Linux-v4.19/include/linux/ |
D | seqlock.h | 133 smp_rmb(); in raw_read_seqcount() 149 smp_rmb(); in raw_read_seqcount_begin() 185 smp_rmb(); in raw_seqcount_begin() 220 smp_rmb(); in read_seqcount_retry()
|
D | cnt32_to_63.h | 100 smp_rmb(); \
|
/Linux-v4.19/include/drm/ |
D | spsc_queue.h | 94 smp_rmb(); in spsc_queue_pop() 111 smp_rmb(); in spsc_queue_pop()
|
/Linux-v4.19/drivers/staging/vc04_services/interface/vchiq_arm/ |
D | vchiq_util.c | 119 smp_rmb(); in vchiu_queue_peek() 137 smp_rmb(); in vchiu_queue_pop()
|
/Linux-v4.19/arch/mips/kernel/ |
D | rtlx.c | 125 smp_rmb(); in rtlx_open() 134 smp_rmb(); in rtlx_open() 254 smp_rmb(); in rtlx_read() 297 smp_rmb(); in rtlx_write()
|
/Linux-v4.19/Documentation/RCU/ |
D | rculist_nulls.txt | 34 but a version with an additional memory barrier (smp_rmb()) 40 pos && ({ next = pos->next; smp_rmb(); prefetch(next); 1; }) && 47 And note the traditional hlist_for_each_entry_rcu() misses this smp_rmb() : 112 With hlist_nulls we can avoid extra smp_rmb() in lockless_lookup()
|
/Linux-v4.19/arch/powerpc/kvm/ |
D | irq.h | 18 smp_rmb(); in irqchip_in_kernel()
|
/Linux-v4.19/arch/powerpc/include/asm/book3s/64/ |
D | hash-64k.h | 78 smp_rmb(); in __real_pte() 184 smp_rmb(); in get_hpte_slot_array()
|
/Linux-v4.19/kernel/livepatch/ |
D | patch.c | 91 smp_rmb(); in klp_ftrace_handler() 101 smp_rmb(); in klp_ftrace_handler()
|
/Linux-v4.19/drivers/cpuidle/ |
D | coupled.c | 312 smp_rmb(); in cpuidle_coupled_get_state() 500 smp_rmb(); in cpuidle_enter_state_coupled() 581 smp_rmb(); in cpuidle_enter_state_coupled()
|
/Linux-v4.19/arch/arm/plat-versatile/ |
D | platsmp.c | 79 smp_rmb(); in versatile_boot_secondary()
|
/Linux-v4.19/lib/ |
D | is_single_threaded.c | 50 smp_rmb(); in current_is_single_threaded()
|
/Linux-v4.19/arch/sh/mm/ |
D | gup.c | 60 smp_rmb(); in gup_get_pte() 62 smp_rmb(); in gup_get_pte()
|
/Linux-v4.19/drivers/firmware/tegra/ |
D | ivc.c | 290 smp_rmb(); in tegra_ivc_read_get_next_frame() 440 smp_rmb(); in tegra_ivc_notified() 479 smp_rmb(); in tegra_ivc_notified() 519 smp_rmb(); in tegra_ivc_notified()
|
/Linux-v4.19/drivers/gpu/host1x/ |
D | syncpt.c | 323 smp_rmb(); in host1x_syncpt_is_expired() 487 smp_rmb(); in host1x_syncpt_read_max() 502 smp_rmb(); in host1x_syncpt_read_min()
|
/Linux-v4.19/arch/arm/vdso/ |
D | vgettimeofday.c | 52 smp_rmb(); /* Pairs with smp_wmb in vdso_write_end */ in vdso_read_begin() 58 smp_rmb(); /* Pairs with smp_wmb in vdso_write_begin */ in vdso_read_retry()
|
/Linux-v4.19/kernel/printk/ |
D | printk_safe.c | 102 smp_rmb(); in printk_safe_log_store() 232 smp_rmb(); in __printk_safe_flush()
|
/Linux-v4.19/tools/memory-model/Documentation/ |
D | recipes.txt | 291 and to use smp_load_acquire() instead of smp_rmb(). However, the older 292 smp_wmb() and smp_rmb() APIs are still heavily used, so it is important 306 smp_rmb(); 311 smp_rmb() macro orders prior loads against later loads. Therefore, if 326 smp_rmb(); 335 * (A) smp_rmb() (C) 342 write side and smp_rmb() on the read side. 345 or smp_rmb(), any code fragment that would work with smp_rmb() and 390 * (A) smp_rmb() (C)
|