Searched refs:smp_load_acquire (Results 1 – 25 of 64) sorted by relevance
123
| /Linux-v4.19/include/asm-generic/ |
| D | barrier.h | 160 #ifndef smp_load_acquire 161 #define smp_load_acquire(p) __smp_load_acquire(p) macro 187 #ifndef smp_load_acquire 188 #define smp_load_acquire(p) \ macro
|
| /Linux-v4.19/tools/memory-model/litmus-tests/ |
| D | MP+pooncerelease+poacquireonce.litmus | 7 * smp_load_acquire() provide sufficient ordering for the message-passing 24 r0 = smp_load_acquire(y);
|
| D | MP+polockonce+poacquiresilsil.litmus | 9 * the smp_load_acquire() executed before the lock was acquired (loosely 29 r1 = smp_load_acquire(x);
|
| D | MP+polockmbonce+poacquiresilsil.litmus | 9 * returns false and the second true, we know that the smp_load_acquire() 30 r1 = smp_load_acquire(x);
|
| D | ISA2+pooncerelease+poacquirerelease+poacquireonce.litmus | 26 r0 = smp_load_acquire(y); 35 r0 = smp_load_acquire(z);
|
| D | S+fencewmbonceonce+poacquireonce.litmus | 23 r0 = smp_load_acquire(y);
|
| D | LB+poacquireonce+pooncerelease.litmus | 25 r0 = smp_load_acquire(y);
|
| D | MP+fencewmbonceonce+fencermbonceonce.litmus | 8 * is usually better to use smp_store_release() and smp_load_acquire().
|
| D | S+poonceonces.litmus | 8 * is replaced by WRITE_ONCE() and the smp_load_acquire() replaced by
|
| D | ISA2+poonceonces.litmus | 9 * of the smp_load_acquire() invocations are replaced by READ_ONCE()?
|
| D | MP+polocks.litmus | 7 * stand in for smp_load_acquire() and smp_store_release(), respectively.
|
| D | MP+porevlocks.litmus | 7 * stand in for smp_load_acquire() and smp_store_release(), respectively.
|
| D | Z6.0+pooncerelease+poacquirerelease+fencembonceonce.litmus | 29 r0 = smp_load_acquire(y);
|
| /Linux-v4.19/lib/ |
| D | stackdepot.c | 91 if (smp_load_acquire(&next_slab_inited)) in init_stack_slab() 238 found = find_stack(smp_load_acquire(bucket), trace->entries, in depot_save_stack() 251 if (unlikely(!smp_load_acquire(&next_slab_inited))) { in depot_save_stack()
|
| D | llist.c | 69 entry = smp_load_acquire(&head->first); in llist_del_first()
|
| /Linux-v4.19/arch/arm/include/asm/ |
| D | mcs_spinlock.h | 13 while (!(smp_load_acquire(lock))) \
|
| /Linux-v4.19/drivers/media/dvb-core/ |
| D | dvb_ringbuffer.c | 60 return (rbuf->pread == smp_load_acquire(&rbuf->pwrite)); in dvb_ringbuffer_empty() 90 avail = smp_load_acquire(&rbuf->pwrite) - rbuf->pread; in dvb_ringbuffer_avail() 106 smp_store_release(&rbuf->pread, smp_load_acquire(&rbuf->pwrite)); in dvb_ringbuffer_flush()
|
| /Linux-v4.19/tools/memory-model/Documentation/ |
| D | recipes.txt | 216 Use of smp_store_release() and smp_load_acquire() is one way to force 228 r0 = smp_load_acquire(&y); 233 store, while the smp_load_acquire macro orders the load against any 246 use of smp_store_release() and smp_load_acquire(), except that both 291 and to use smp_load_acquire() instead of smp_rmb(). However, the older 421 r0 = smp_load_acquire(y); 427 r1 = smp_load_acquire(z); 435 example, ordering would still be preserved if CPU1()'s smp_load_acquire() 450 r0 = smp_load_acquire(y);
|
| /Linux-v4.19/drivers/tty/ |
| D | tty_buffer.c | 226 while ((next = smp_load_acquire(&buf->head->next)) != NULL) { in tty_buffer_flush() 511 next = smp_load_acquire(&head->next); in flush_to_ldisc() 515 count = smp_load_acquire(&head->commit) - head->read; in flush_to_ldisc()
|
| /Linux-v4.19/kernel/locking/ |
| D | percpu-rwsem.c | 64 if (likely(!smp_load_acquire(&sem->readers_block))) in __percpu_down_read()
|
| D | osq_lock.c | 169 if (smp_load_acquire(&node->locked)) in osq_lock()
|
| /Linux-v4.19/fs/ |
| D | locks.c | 212 ctx = smp_load_acquire(&inode->i_flctx); in locks_get_lock_context() 231 ctx = smp_load_acquire(&inode->i_flctx); in locks_get_lock_context() 819 ctx = smp_load_acquire(&inode->i_flctx); in posix_test_lock() 1265 ctx = smp_load_acquire(&inode->i_flctx); in locks_mandatory_locked() 1459 ctx = smp_load_acquire(&inode->i_flctx); in __break_lease() 1563 ctx = smp_load_acquire(&inode->i_flctx); in lease_get_mtime() 1610 ctx = smp_load_acquire(&inode->i_flctx); in fcntl_getlease() 1784 ctx = smp_load_acquire(&inode->i_flctx); in generic_delete_lease() 2475 ctx = smp_load_acquire(&inode->i_flctx); in locks_remove_posix() 2552 ctx = smp_load_acquire(&locks_inode(filp)->i_flctx); in locks_remove_file() [all …]
|
| /Linux-v4.19/tools/memory-model/ |
| D | linux-kernel.def | 14 smp_load_acquire(X) __load{acquire}(*X) 52 atomic_read_acquire(X) smp_load_acquire(X)
|
| /Linux-v4.19/arch/arm64/include/asm/ |
| D | barrier.h | 149 VAL = smp_load_acquire(__PTR); \
|
| /Linux-v4.19/net/rxrpc/ |
| D | call_accept.c | 263 call_head = smp_load_acquire(&b->call_backlog_head); in rxrpc_alloc_incoming_call() 266 conn_head = smp_load_acquire(&b->conn_backlog_head); in rxrpc_alloc_incoming_call() 270 peer_head = smp_load_acquire(&b->peer_backlog_head); in rxrpc_alloc_incoming_call()
|
123