/Linux-v4.19/include/asm-generic/ |
D | atomic-instrumented.h | 19 static __always_inline int atomic_read(const atomic_t *v) in atomic_read() 25 static __always_inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() 31 static __always_inline void atomic_set(atomic_t *v, int i) in atomic_set() 37 static __always_inline void atomic64_set(atomic64_t *v, s64 i) in atomic64_set() 43 static __always_inline int atomic_xchg(atomic_t *v, int i) in atomic_xchg() 49 static __always_inline s64 atomic64_xchg(atomic64_t *v, s64 i) in atomic64_xchg() 55 static __always_inline int atomic_cmpxchg(atomic_t *v, int old, int new) in atomic_cmpxchg() 61 static __always_inline s64 atomic64_cmpxchg(atomic64_t *v, s64 old, s64 new) in atomic64_cmpxchg() 69 static __always_inline bool atomic_try_cmpxchg(atomic_t *v, int *old, int new) in atomic_try_cmpxchg() 79 static __always_inline bool atomic64_try_cmpxchg(atomic64_t *v, s64 *old, s64 new) in atomic64_try_cmpxchg() [all …]
|
D | preempt.h | 9 static __always_inline int preempt_count(void) in preempt_count() 14 static __always_inline volatile int *preempt_count_ptr(void) in preempt_count_ptr() 19 static __always_inline void preempt_count_set(int pc) in preempt_count_set() 35 static __always_inline void set_preempt_need_resched(void) in set_preempt_need_resched() 39 static __always_inline void clear_preempt_need_resched(void) in clear_preempt_need_resched() 43 static __always_inline bool test_preempt_need_resched(void) in test_preempt_need_resched() 52 static __always_inline void __preempt_count_add(int val) in __preempt_count_add() 57 static __always_inline void __preempt_count_sub(int val) in __preempt_count_sub() 62 static __always_inline bool __preempt_count_dec_and_test(void) in __preempt_count_dec_and_test() 75 static __always_inline bool should_resched(int preempt_offset) in should_resched()
|
D | qspinlock.h | 29 static __always_inline int queued_spin_is_locked(struct qspinlock *lock) in queued_spin_is_locked() 48 static __always_inline int queued_spin_value_unlocked(struct qspinlock lock) in queued_spin_value_unlocked() 58 static __always_inline int queued_spin_is_contended(struct qspinlock *lock) in queued_spin_is_contended() 67 static __always_inline int queued_spin_trylock(struct qspinlock *lock) in queued_spin_trylock() 81 static __always_inline void queued_spin_lock(struct qspinlock *lock) in queued_spin_lock() 96 static __always_inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock() 106 static __always_inline bool virt_spin_lock(struct qspinlock *lock) in virt_spin_lock()
|
/Linux-v4.19/arch/x86/include/asm/ |
D | bitops.h | 72 static __always_inline void 95 static __always_inline void __set_bit(long nr, volatile unsigned long *addr) in __set_bit() 110 static __always_inline void 132 static __always_inline void clear_bit_unlock(long nr, volatile unsigned long *addr) in clear_bit_unlock() 138 static __always_inline void __clear_bit(long nr, volatile unsigned long *addr) in __clear_bit() 143 static __always_inline bool clear_bit_unlock_is_negative_byte(long nr, volatile unsigned long *addr) in clear_bit_unlock_is_negative_byte() 168 static __always_inline void __clear_bit_unlock(long nr, volatile unsigned long *addr) in __clear_bit_unlock() 183 static __always_inline void __change_bit(long nr, volatile unsigned long *addr) in __change_bit() 197 static __always_inline void change_bit(long nr, volatile unsigned long *addr) in change_bit() 218 static __always_inline bool test_and_set_bit(long nr, volatile unsigned long *addr) in test_and_set_bit() [all …]
|
D | atomic.h | 25 static __always_inline int arch_atomic_read(const atomic_t *v) in arch_atomic_read() 41 static __always_inline void arch_atomic_set(atomic_t *v, int i) in arch_atomic_set() 53 static __always_inline void arch_atomic_add(int i, atomic_t *v) in arch_atomic_add() 67 static __always_inline void arch_atomic_sub(int i, atomic_t *v) in arch_atomic_sub() 83 static __always_inline bool arch_atomic_sub_and_test(int i, atomic_t *v) in arch_atomic_sub_and_test() 95 static __always_inline void arch_atomic_inc(atomic_t *v) in arch_atomic_inc() 108 static __always_inline void arch_atomic_dec(atomic_t *v) in arch_atomic_dec() 123 static __always_inline bool arch_atomic_dec_and_test(atomic_t *v) in arch_atomic_dec_and_test() 137 static __always_inline bool arch_atomic_inc_and_test(atomic_t *v) in arch_atomic_inc_and_test() 152 static __always_inline bool arch_atomic_add_negative(int i, atomic_t *v) in arch_atomic_add_negative() [all …]
|
D | preempt.h | 21 static __always_inline int preempt_count(void) in preempt_count() 26 static __always_inline void preempt_count_set(int pc) in preempt_count_set() 55 static __always_inline void set_preempt_need_resched(void) in set_preempt_need_resched() 60 static __always_inline void clear_preempt_need_resched(void) in clear_preempt_need_resched() 65 static __always_inline bool test_preempt_need_resched(void) in test_preempt_need_resched() 74 static __always_inline void __preempt_count_add(int val) in __preempt_count_add() 79 static __always_inline void __preempt_count_sub(int val) in __preempt_count_sub() 89 static __always_inline bool __preempt_count_dec_and_test(void) in __preempt_count_dec_and_test() 97 static __always_inline bool should_resched(int preempt_offset) in should_resched()
|
D | refcount.h | 42 static __always_inline void refcount_add(unsigned int i, refcount_t *r) in refcount_add() 51 static __always_inline void refcount_inc(refcount_t *r) in refcount_inc() 59 static __always_inline void refcount_dec(refcount_t *r) in refcount_dec() 67 static __always_inline __must_check 74 static __always_inline __must_check bool refcount_dec_and_test(refcount_t *r) in refcount_dec_and_test() 80 static __always_inline __must_check 105 static __always_inline __must_check bool refcount_inc_not_zero(refcount_t *r) in refcount_inc_not_zero()
|
/Linux-v4.19/include/linux/unaligned/ |
D | access_ok.h | 8 static __always_inline u16 get_unaligned_le16(const void *p) in get_unaligned_le16() 13 static __always_inline u32 get_unaligned_le32(const void *p) in get_unaligned_le32() 18 static __always_inline u64 get_unaligned_le64(const void *p) in get_unaligned_le64() 23 static __always_inline u16 get_unaligned_be16(const void *p) in get_unaligned_be16() 28 static __always_inline u32 get_unaligned_be32(const void *p) in get_unaligned_be32() 33 static __always_inline u64 get_unaligned_be64(const void *p) in get_unaligned_be64() 38 static __always_inline void put_unaligned_le16(u16 val, void *p) in put_unaligned_le16() 43 static __always_inline void put_unaligned_le32(u32 val, void *p) in put_unaligned_le32() 48 static __always_inline void put_unaligned_le64(u64 val, void *p) in put_unaligned_le64() 53 static __always_inline void put_unaligned_be16(u16 val, void *p) in put_unaligned_be16() [all …]
|
/Linux-v4.19/arch/riscv/include/asm/ |
D | atomic.h | 34 static __always_inline int atomic_read(const atomic_t *v) in atomic_read() 38 static __always_inline void atomic_set(atomic_t *v, int i) in atomic_set() 45 static __always_inline long atomic64_read(const atomic64_t *v) in atomic64_read() 49 static __always_inline void atomic64_set(atomic64_t *v, long i) in atomic64_set() 61 static __always_inline \ 95 static __always_inline \ in ATOMIC_OPS() 107 static __always_inline \ 120 static __always_inline \ 126 static __always_inline \ 206 static __always_inline int atomic_fetch_add_unless(atomic_t *v, int a, int u) [all …]
|
/Linux-v4.19/include/uapi/linux/byteorder/ |
D | big_endian.h | 44 static __always_inline __le64 __cpu_to_le64p(const __u64 *p) in __cpu_to_le64p() 48 static __always_inline __u64 __le64_to_cpup(const __le64 *p) in __le64_to_cpup() 52 static __always_inline __le32 __cpu_to_le32p(const __u32 *p) in __cpu_to_le32p() 56 static __always_inline __u32 __le32_to_cpup(const __le32 *p) in __le32_to_cpup() 60 static __always_inline __le16 __cpu_to_le16p(const __u16 *p) in __cpu_to_le16p() 64 static __always_inline __u16 __le16_to_cpup(const __le16 *p) in __le16_to_cpup() 68 static __always_inline __be64 __cpu_to_be64p(const __u64 *p) in __cpu_to_be64p() 72 static __always_inline __u64 __be64_to_cpup(const __be64 *p) in __be64_to_cpup() 76 static __always_inline __be32 __cpu_to_be32p(const __u32 *p) in __cpu_to_be32p() 80 static __always_inline __u32 __be32_to_cpup(const __be32 *p) in __be32_to_cpup() [all …]
|
D | little_endian.h | 44 static __always_inline __le64 __cpu_to_le64p(const __u64 *p) in __cpu_to_le64p() 48 static __always_inline __u64 __le64_to_cpup(const __le64 *p) in __le64_to_cpup() 52 static __always_inline __le32 __cpu_to_le32p(const __u32 *p) in __cpu_to_le32p() 56 static __always_inline __u32 __le32_to_cpup(const __le32 *p) in __le32_to_cpup() 60 static __always_inline __le16 __cpu_to_le16p(const __u16 *p) in __cpu_to_le16p() 64 static __always_inline __u16 __le16_to_cpup(const __le16 *p) in __le16_to_cpup() 68 static __always_inline __be64 __cpu_to_be64p(const __u64 *p) in __cpu_to_be64p() 72 static __always_inline __u64 __be64_to_cpup(const __be64 *p) in __be64_to_cpup() 76 static __always_inline __be32 __cpu_to_be32p(const __u32 *p) in __cpu_to_be32p() 80 static __always_inline __u32 __be32_to_cpup(const __be32 *p) in __be32_to_cpup() [all …]
|
/Linux-v4.19/arch/powerpc/include/asm/ |
D | cmpxchg.h | 88 static __always_inline unsigned long 105 static __always_inline unsigned long 123 static __always_inline unsigned long 140 static __always_inline unsigned long 158 static __always_inline unsigned long 177 static __always_inline unsigned long 222 static __always_inline unsigned long 245 static __always_inline unsigned long 267 static __always_inline unsigned long 295 static __always_inline unsigned long [all …]
|
/Linux-v4.19/kernel/locking/ |
D | qspinlock.c | 141 static __always_inline void clear_pending(struct qspinlock *lock) in clear_pending() 154 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) in clear_pending_set_locked() 169 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 187 static __always_inline void clear_pending(struct qspinlock *lock) in clear_pending() 198 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) in clear_pending_set_locked() 213 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 240 static __always_inline void set_locked(struct qspinlock *lock) in set_locked() 251 static __always_inline void __pv_init_node(struct mcs_spinlock *node) { } in __pv_init_node() 252 static __always_inline void __pv_wait_node(struct mcs_spinlock *node, in __pv_wait_node() 254 static __always_inline void __pv_kick_node(struct qspinlock *lock, in __pv_kick_node() [all …]
|
/Linux-v4.19/arch/ia64/include/asm/ |
D | spinlock.h | 43 static __always_inline void __ticket_spin_lock(arch_spinlock_t *lock) in __ticket_spin_lock() 63 static __always_inline int __ticket_spin_trylock(arch_spinlock_t *lock) in __ticket_spin_trylock() 72 static __always_inline void __ticket_spin_unlock(arch_spinlock_t *lock) in __ticket_spin_unlock() 94 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) in arch_spin_value_unlocked() 110 static __always_inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() 115 static __always_inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock() 120 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() 125 static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock, in arch_spin_lock_flags() 134 static __always_inline void 186 static __always_inline void
|
/Linux-v4.19/include/linux/ |
D | spinlock.h | 316 static __always_inline raw_spinlock_t *spinlock_check(spinlock_t *lock) in spinlock_check() 327 static __always_inline void spin_lock(spinlock_t *lock) in spin_lock() 332 static __always_inline void spin_lock_bh(spinlock_t *lock) in spin_lock_bh() 337 static __always_inline int spin_trylock(spinlock_t *lock) in spin_trylock() 352 static __always_inline void spin_lock_irq(spinlock_t *lock) in spin_lock_irq() 367 static __always_inline void spin_unlock(spinlock_t *lock) in spin_unlock() 372 static __always_inline void spin_unlock_bh(spinlock_t *lock) in spin_unlock_bh() 377 static __always_inline void spin_unlock_irq(spinlock_t *lock) in spin_unlock_irq() 382 static __always_inline void spin_unlock_irqrestore(spinlock_t *lock, unsigned long flags) in spin_unlock_irqrestore() 387 static __always_inline int spin_trylock_bh(spinlock_t *lock) in spin_trylock_bh() [all …]
|
D | mm_inline.h | 26 static __always_inline void __update_lru_size(struct lruvec *lruvec, in __update_lru_size() 37 static __always_inline void update_lru_size(struct lruvec *lruvec, in update_lru_size() 47 static __always_inline void add_page_to_lru_list(struct page *page, in add_page_to_lru_list() 54 static __always_inline void add_page_to_lru_list_tail(struct page *page, in add_page_to_lru_list_tail() 61 static __always_inline void del_page_from_lru_list(struct page *page, in del_page_from_lru_list() 90 static __always_inline enum lru_list page_off_lru(struct page *page) in page_off_lru() 114 static __always_inline enum lru_list page_lru(struct page *page) in page_lru()
|
D | uaccess.h | 61 static __always_inline unsigned long 69 static __always_inline unsigned long 91 static __always_inline unsigned long 99 static __always_inline unsigned long 143 static __always_inline unsigned long __must_check 151 static __always_inline unsigned long __must_check 159 static __always_inline unsigned long __must_check 169 static __always_inline void pagefault_disabled_inc(void) in pagefault_disabled_inc() 174 static __always_inline void pagefault_disabled_dec(void) in pagefault_disabled_dec()
|
D | page-flags.h | 149 static __always_inline int PageTail(struct page *page) in PageTail() 154 static __always_inline int PageCompound(struct page *page) in PageCompound() 207 static __always_inline int Page##uname(struct page *page) \ 211 static __always_inline void SetPage##uname(struct page *page) \ 215 static __always_inline void ClearPage##uname(struct page *page) \ 219 static __always_inline void __SetPage##uname(struct page *page) \ 223 static __always_inline void __ClearPage##uname(struct page *page) \ 227 static __always_inline int TestSetPage##uname(struct page *page) \ 231 static __always_inline int TestClearPage##uname(struct page *page) \ 335 static __always_inline int PageSwapCache(struct page *page) in PAGEFLAG() [all …]
|
D | rbtree_latch.h | 69 static __always_inline struct latch_tree_node * 75 static __always_inline void 99 static __always_inline void 105 static __always_inline struct latch_tree_node * 143 static __always_inline void 170 static __always_inline void 199 static __always_inline struct latch_tree_node *
|
D | bitfield.h | 111 static __always_inline u64 field_multiplier(u64 field) in field_multiplier() 117 static __always_inline u64 field_mask(u64 field) in field_mask() 122 static __always_inline __##type type##_encode_bits(base v, base field) \ 128 static __always_inline __##type type##_replace_bits(__##type old, \ 133 static __always_inline void type##p_replace_bits(__##type *p, \ 138 static __always_inline base type##_get_bits(__##type v, base field) \
|
D | slab.h | 312 static __always_inline unsigned int kmalloc_index(size_t size) in kmalloc_index() 373 static __always_inline void kfree_bulk(size_t size, void **p) in kfree_bulk() 382 static __always_inline void *__kmalloc_node(size_t size, gfp_t flags, int node) in __kmalloc_node() 387 static __always_inline void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t flags, int node) in kmem_cache_alloc_node() 401 static __always_inline void * 411 static __always_inline void *kmem_cache_alloc_trace(struct kmem_cache *s, in kmem_cache_alloc_trace() 420 static __always_inline void * 437 static __always_inline void * 444 static __always_inline void *kmalloc_large(size_t size, gfp_t flags) in kmalloc_large() 501 static __always_inline void *kmalloc(size_t size, gfp_t flags) in kmalloc() [all …]
|
/Linux-v4.19/arch/mips/vdso/ |
D | gettimeofday.c | 23 static __always_inline long gettimeofday_fallback(struct timeval *_tv, in gettimeofday_fallback() 44 static __always_inline long clock_gettime_fallback(clockid_t _clkid, in clock_gettime_fallback() 63 static __always_inline int do_realtime_coarse(struct timespec *ts, in do_realtime_coarse() 78 static __always_inline int do_monotonic_coarse(struct timespec *ts, in do_monotonic_coarse() 103 static __always_inline u64 read_r4k_count(void) in read_r4k_count() 121 static __always_inline u64 read_gic_count(const union mips_vdso_data *data) in read_gic_count() 137 static __always_inline u64 get_ns(const union mips_vdso_data *data) in get_ns() 164 static __always_inline int do_realtime(struct timespec *ts, in do_realtime() 186 static __always_inline int do_monotonic(struct timespec *ts, in do_monotonic()
|
/Linux-v4.19/drivers/misc/mic/scif/ |
D | scif_map.h | 23 static __always_inline void * 43 static __always_inline void 57 static __always_inline int 78 static __always_inline void 90 static __always_inline void * 104 static __always_inline void 114 static __always_inline int
|
/Linux-v4.19/tools/include/linux/ |
D | compiler.h | 17 #ifndef __always_inline 18 # define __always_inline inline __attribute__((always_inline)) macro 37 #undef __always_inline 38 #define __always_inline inline macro 105 static __always_inline void __read_once_size(const volatile void *p, void *res, int size) in __read_once_size() 119 static __always_inline void __write_once_size(volatile void *p, void *res, int size) in __write_once_size()
|
/Linux-v4.19/tools/testing/selftests/x86/ |
D | mpx-mini-test.c | 43 #ifndef __always_inline 44 #define __always_inline inline __attribute__((always_inline) macro 181 static __always_inline void xrstor_state(struct xsave_struct *fx, uint64_t mask) in xrstor_state() 191 static __always_inline void xsave_state_1(void *_fx, uint64_t mask) in xsave_state_1() 782 static __always_inline void xsave_state(void *_fx, uint64_t mask) in xsave_state() 793 static __always_inline void mpx_clear_bnd0(void) in mpx_clear_bnd0() 804 static __always_inline void mpx_make_bound_helper(unsigned long ptr, in mpx_make_bound_helper() 814 static __always_inline void mpx_check_lowerbound_helper(unsigned long ptr) in mpx_check_lowerbound_helper() 823 static __always_inline void mpx_check_upperbound_helper(unsigned long ptr) in mpx_check_upperbound_helper() 832 static __always_inline void mpx_movbndreg_helper() in mpx_movbndreg_helper() [all …]
|