| /Linux-v5.4/arch/arc/include/asm/ |
| D | spinlock.h | 40 smp_mb(); in arch_spin_lock() 62 smp_mb(); in arch_spin_trylock() 69 smp_mb(); in arch_spin_unlock() 105 smp_mb(); in arch_read_lock() 129 smp_mb(); in arch_read_trylock() 163 smp_mb(); in arch_write_lock() 188 smp_mb(); in arch_write_trylock() 197 smp_mb(); in arch_read_unlock() 215 smp_mb(); in arch_write_unlock() 231 smp_mb(); in arch_spin_lock() [all …]
|
| D | atomic.h | 52 smp_mb(); \ 64 smp_mb(); \ 78 smp_mb(); \ 91 smp_mb(); \ 241 smp_mb(); \ 252 smp_mb(); \ 265 smp_mb(); \ 276 smp_mb(); \ 383 smp_mb(); \ 396 smp_mb(); \ [all …]
|
| D | cmpxchg.h | 25 smp_mb(); in __cmpxchg() 39 smp_mb(); in __cmpxchg() 72 smp_mb(); in __cmpxchg() 85 smp_mb(); in __cmpxchg() 119 smp_mb(); in __xchg() 127 smp_mb(); in __xchg() 183 smp_mb(); in __xchg() 194 smp_mb(); in __xchg()
|
| D | futex.h | 20 smp_mb(); \ 42 smp_mb() \ 48 smp_mb(); \ 69 smp_mb() \ 132 smp_mb(); in futex_atomic_cmpxchg_inatomic() 159 smp_mb(); in futex_atomic_cmpxchg_inatomic()
|
| D | bitops.h | 72 smp_mb(); \ 83 smp_mb(); \ 171 smp_mb(); \ 182 smp_mb(); \
|
| /Linux-v5.4/arch/csky/include/asm/ |
| D | spinlock.h | 34 smp_mb(); in arch_spin_lock() 60 smp_mb(); in arch_spin_trylock() 67 smp_mb(); in arch_spin_unlock() 92 #define smp_mb__after_spinlock() smp_mb() 113 smp_mb(); in arch_spin_lock() 118 smp_mb(); in arch_spin_unlock() 140 smp_mb(); in arch_spin_trylock() 164 smp_mb(); in arch_read_lock() 172 smp_mb(); in arch_read_unlock() 201 smp_mb(); in arch_read_trylock() [all …]
|
| D | atomic.h | 17 smp_mb(); in __atomic_add_unless() 33 smp_mb(); in __atomic_add_unless() 58 smp_mb(); \ 68 smp_mb(); \ 78 smp_mb(); \ 88 smp_mb(); \
|
| D | cmpxchg.h | 19 smp_mb(); \ 28 smp_mb(); \ 47 smp_mb(); \ 59 smp_mb(); \
|
| /Linux-v5.4/arch/arm/include/asm/ |
| D | spinlock.h | 78 smp_mb(); in arch_spin_lock() 100 smp_mb(); in arch_spin_trylock() 109 smp_mb(); in arch_spin_unlock() 155 smp_mb(); in arch_write_lock() 175 smp_mb(); in arch_write_trylock() 184 smp_mb(); in arch_write_unlock() 224 smp_mb(); in arch_read_lock() 231 smp_mb(); in arch_read_unlock() 266 smp_mb(); in arch_read_trylock()
|
| D | futex.h | 28 smp_mb(); \ 56 smp_mb(); in futex_atomic_cmpxchg_inatomic() 73 smp_mb(); in futex_atomic_cmpxchg_inatomic()
|
| /Linux-v5.4/tools/testing/selftests/rcutorture/formal/srcu-cbmc/src/ |
| D | barriers.h | 8 #define smp_mb() __sync_synchronize() macro 15 #define smp_mb() __CPROVER_fence("WWfence", "RRfence", "RWfence", "WRfence", \ macro 27 #define sync_smp_mb() smp_mb() 33 #define rs_smp_mb() smp_mb()
|
| /Linux-v5.4/kernel/sched/ |
| D | membarrier.c | 30 smp_mb(); /* IPIs should be serializing but paranoid. */ in ipi_mb() 47 smp_mb(); in ipi_sync_rq_state() 57 smp_mb(); in membarrier_exec_mmap() 78 smp_mb(); /* system call entry is not a mb. */ in membarrier_global_expedited() 128 smp_mb(); /* exit from system call is not a mb */ in membarrier_global_expedited() 157 smp_mb(); /* system call entry is not a mb. */ in membarrier_private_expedited() 195 smp_mb(); /* exit from system call is not a mb */ in membarrier_private_expedited() 216 smp_mb(); in sync_runqueues_membarrier_state()
|
| /Linux-v5.4/tools/include/asm/ |
| D | barrier.h | 46 #ifndef smp_mb 47 # define smp_mb() mb() macro 53 smp_mb(); \ 62 smp_mb(); \
|
| /Linux-v5.4/tools/memory-model/litmus-tests/ |
| D | IRIW+fencembonceonces+OnceOnce.litmus | 6 * Test of independent reads from independent writes with smp_mb() 7 * between each pairs of reads. In other words, is smp_mb() sufficient to 26 smp_mb(); 41 smp_mb();
|
| D | R+fencembonceonces.litmus | 6 * This is the fully ordered (via smp_mb()) version of one of the classic 17 smp_mb(); 26 smp_mb();
|
| /Linux-v5.4/arch/alpha/include/asm/ |
| D | cmpxchg.h | 49 smp_mb(); \ 52 smp_mb(); \ 61 smp_mb(); \ 64 smp_mb(); \
|
| D | atomic.h | 220 smp_mb(); in ATOMIC_OPS() 235 smp_mb(); in ATOMIC_OPS() 252 smp_mb(); in atomic64_fetch_add_unless() 267 smp_mb(); in atomic64_fetch_add_unless() 282 smp_mb(); in atomic64_dec_if_positive() 296 smp_mb(); in atomic64_dec_if_positive()
|
| /Linux-v5.4/kernel/locking/ |
| D | percpu-rwsem.c | 61 smp_mb(); /* A matches D */ in __percpu_down_read() 99 smp_mb(); /* B matches C */ in __percpu_up_read() 138 smp_mb(); /* C matches B */ in readers_active_check() 156 smp_mb(); /* D matches A */ in percpu_down_write()
|
| /Linux-v5.4/arch/sh/kernel/ |
| D | ftrace.c | 137 smp_mb(); in arch_ftrace_nmi_enter() 143 smp_mb(); in arch_ftrace_nmi_exit() 174 smp_mb(); in do_ftrace_mod_code() 179 smp_mb(); in do_ftrace_mod_code() 184 smp_mb(); in do_ftrace_mod_code()
|
| /Linux-v5.4/arch/nds32/include/asm/ |
| D | futex.h | 23 smp_mb(); \ 46 smp_mb(); in futex_atomic_cmpxchg_inatomic() 58 smp_mb(); in futex_atomic_cmpxchg_inatomic()
|
| /Linux-v5.4/include/asm-generic/ |
| D | barrier.h | 71 #ifndef smp_mb 72 #define smp_mb() __smp_mb() macro 89 #ifndef smp_mb 90 #define smp_mb() barrier() macro
|
| /Linux-v5.4/tools/virtio/ringtest/ |
| D | ring.c | 183 smp_mb(); in enable_call() 193 smp_mb(); in kick_available() 215 smp_mb(); in enable_kick() 260 smp_mb(); in call_used()
|
| /Linux-v5.4/arch/arm64/include/asm/vdso/ |
| D | compat_barrier.h | 34 #undef smp_mb 38 #define smp_mb() aarch32_smp_mb() macro
|
| /Linux-v5.4/kernel/rcu/ |
| D | rcu_segcblist.c | 114 smp_mb(); /* Up to the caller! */ in rcu_segcblist_add_len() 116 smp_mb(); /* Up to the caller! */ in rcu_segcblist_add_len() 144 smp_mb(); /* Up to the caller! */ in rcu_segcblist_xchg_len() 146 smp_mb(); /* Up to the caller! */ in rcu_segcblist_xchg_len() 261 smp_mb(); /* Ensure counts are updated before callback is enqueued. */ in rcu_segcblist_enqueue() 287 smp_mb(); /* Ensure counts are updated before callback is entrained. */ in rcu_segcblist_entrain()
|
| /Linux-v5.4/include/linux/ |
| D | freezer.h | 127 smp_mb(); in freezer_count() 135 smp_mb(); in freezer_count_unsafe() 158 smp_mb(); in freezer_should_skip()
|