/Linux-v5.4/arch/parisc/include/asm/ |
D | spinlock.h | 35 static inline void arch_spin_unlock(arch_spinlock_t *x) in arch_spin_unlock() function 79 arch_spin_unlock(&rw->lock); in arch_read_lock() 91 arch_spin_unlock(&rw->lock); in arch_read_unlock() 104 arch_spin_unlock(&rw->lock); in arch_read_trylock() 131 arch_spin_unlock(&rw->lock); in arch_write_lock() 148 arch_spin_unlock(&rw->lock); in arch_write_unlock() 165 arch_spin_unlock(&rw->lock); in arch_write_trylock()
|
D | futex.h | 31 arch_spin_unlock(s); in _futex_spin_unlock_irqrestore()
|
/Linux-v5.4/arch/arc/include/asm/ |
D | spinlock.h | 67 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function 267 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function 321 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_trylock() 346 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_trylock() 371 arch_spin_unlock(&(rw->lock_mutex)); in arch_read_unlock() 382 arch_spin_unlock(&(rw->lock_mutex)); in arch_write_unlock()
|
D | smp.h | 116 arch_spin_unlock(&smp_atomic_ops_lock); \ 126 arch_spin_unlock(&smp_bitops_lock); \
|
/Linux-v5.4/kernel/locking/ |
D | qrwlock.c | 54 arch_spin_unlock(&lock->wait_lock); in queued_read_lock_slowpath() 81 arch_spin_unlock(&lock->wait_lock); in queued_write_lock_slowpath()
|
/Linux-v5.4/include/linux/ |
D | spinlock_up.h | 45 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function 66 # define arch_spin_unlock(lock) do { barrier(); (void)(lock); } while (0) macro
|
/Linux-v5.4/arch/arm/common/ |
D | mcpm_entry.c | 232 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_up() 268 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_down() 274 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_power_down() 335 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_suspend() 365 arch_spin_unlock(&mcpm_lock); in mcpm_cpu_powered_up()
|
/Linux-v5.4/arch/csky/include/asm/ |
D | spinlock.h | 65 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function 116 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
|
/Linux-v5.4/kernel/trace/ |
D | trace_stack.c | 282 arch_spin_unlock(&stack_trace_max_lock); in check_stack() 356 arch_spin_unlock(&stack_trace_max_lock); in stack_max_size_write() 406 arch_spin_unlock(&stack_trace_max_lock); in t_stop()
|
D | trace_clock.c | 123 arch_spin_unlock(&trace_clock_struct.lock); in trace_clock_global()
|
/Linux-v5.4/arch/powerpc/kvm/ |
D | book3s_xive.c | 412 arch_spin_unlock(&sb->lock); in xive_lock_and_mask() 467 arch_spin_unlock(&sb->lock); in xive_lock_for_unmask() 710 arch_spin_unlock(&sb->lock); in kvmppc_xive_set_xive() 732 arch_spin_unlock(&sb->lock); in kvmppc_xive_get_xive() 771 arch_spin_unlock(&sb->lock); in kvmppc_xive_int_on() 797 arch_spin_unlock(&sb->lock); in kvmppc_xive_int_off() 987 arch_spin_unlock(&sb->lock); in kvmppc_xive_set_mapped() 1068 arch_spin_unlock(&sb->lock); in kvmppc_xive_clr_mapped() 1105 arch_spin_unlock(&sb->lock); in kvmppc_xive_disable_vcpu_interrupts() 1396 arch_spin_unlock(&sb->lock); in xive_pre_save_mask_irq() [all …]
|
D | book3s_xics.c | 160 arch_spin_unlock(&ics->lock); in write_xive() 216 arch_spin_unlock(&ics->lock); in kvmppc_xics_get_xive() 473 arch_spin_unlock(&ics->lock); in icp_deliver_irq() 502 arch_spin_unlock(&ics->lock); in icp_deliver_irq() 509 arch_spin_unlock(&ics->lock); in icp_deliver_irq() 1009 arch_spin_unlock(&ics->lock); in xics_debug_show() 1226 arch_spin_unlock(&ics->lock); in xics_get_source() 1284 arch_spin_unlock(&ics->lock); in xics_set_source()
|
D | book3s_xive_native.c | 246 arch_spin_unlock(&sb->lock); in xive_native_esb_fault() 385 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_set_source() 433 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_update_source_config() 511 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_sync_source() 825 arch_spin_unlock(&sb->lock); in kvmppc_xive_reset() 907 arch_spin_unlock(&sb->lock); in kvmppc_xive_native_eq_sync()
|
/Linux-v5.4/tools/include/linux/ |
D | spinlock.h | 30 static inline void arch_spin_unlock(arch_spinlock_t *mutex) in arch_spin_unlock() function
|
/Linux-v5.4/arch/hexagon/include/asm/ |
D | spinlock.h | 130 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
|
/Linux-v5.4/arch/x86/kernel/ |
D | tsc_sync.c | 248 arch_spin_unlock(&sync_lock); in check_tsc_warp() 278 arch_spin_unlock(&sync_lock); in check_tsc_warp()
|
/Linux-v5.4/arch/riscv/include/asm/ |
D | spinlock.h | 22 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
|
/Linux-v5.4/arch/alpha/include/asm/ |
D | spinlock.h | 24 static inline void arch_spin_unlock(arch_spinlock_t * lock) in arch_spin_unlock() function
|
/Linux-v5.4/arch/sh/include/asm/ |
D | spinlock-cas.h | 33 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
|
D | spinlock-llsc.h | 46 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
|
/Linux-v5.4/include/asm-generic/ |
D | qspinlock.h | 114 #define arch_spin_unlock(l) queued_spin_unlock(l) macro
|
/Linux-v5.4/arch/s390/lib/ |
D | spinlock.c | 287 arch_spin_unlock(&rw->wait); in arch_read_lock_wait() 310 arch_spin_unlock(&rw->wait); in arch_write_lock_wait()
|
/Linux-v5.4/arch/sparc/include/asm/ |
D | spinlock_32.h | 48 static inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() function
|
/Linux-v5.4/arch/s390/include/asm/ |
D | spinlock.h | 85 static inline void arch_spin_unlock(arch_spinlock_t *lp) in arch_spin_unlock() function
|
/Linux-v5.4/arch/powerpc/platforms/pasemi/ |
D | setup.c | 112 arch_spin_unlock(&timebase_lock); in pas_give_timebase() 128 arch_spin_unlock(&timebase_lock); in pas_take_timebase()
|