| /Linux-v5.4/arch/x86/include/asm/ |
| D | bitops.h | 51 static __always_inline void 65 static __always_inline void 71 static __always_inline void 84 static __always_inline void 91 static __always_inline void 97 static __always_inline bool 110 static __always_inline void 116 static __always_inline void 122 static __always_inline void 135 static __always_inline bool [all …]
|
| D | atomic.h | 25 static __always_inline int arch_atomic_read(const atomic_t *v) in arch_atomic_read() 41 static __always_inline void arch_atomic_set(atomic_t *v, int i) in arch_atomic_set() 53 static __always_inline void arch_atomic_add(int i, atomic_t *v) in arch_atomic_add() 67 static __always_inline void arch_atomic_sub(int i, atomic_t *v) in arch_atomic_sub() 83 static __always_inline bool arch_atomic_sub_and_test(int i, atomic_t *v) in arch_atomic_sub_and_test() 95 static __always_inline void arch_atomic_inc(atomic_t *v) in arch_atomic_inc() 108 static __always_inline void arch_atomic_dec(atomic_t *v) in arch_atomic_dec() 123 static __always_inline bool arch_atomic_dec_and_test(atomic_t *v) in arch_atomic_dec_and_test() 137 static __always_inline bool arch_atomic_inc_and_test(atomic_t *v) in arch_atomic_inc_and_test() 152 static __always_inline bool arch_atomic_add_negative(int i, atomic_t *v) in arch_atomic_add_negative() [all …]
|
| D | preempt.h | 24 static __always_inline int preempt_count(void) in preempt_count() 29 static __always_inline void preempt_count_set(int pc) in preempt_count_set() 58 static __always_inline void set_preempt_need_resched(void) in set_preempt_need_resched() 63 static __always_inline void clear_preempt_need_resched(void) in clear_preempt_need_resched() 68 static __always_inline bool test_preempt_need_resched(void) in test_preempt_need_resched() 77 static __always_inline void __preempt_count_add(int val) in __preempt_count_add() 82 static __always_inline void __preempt_count_sub(int val) in __preempt_count_sub() 92 static __always_inline bool __preempt_count_dec_and_test(void) in __preempt_count_dec_and_test() 100 static __always_inline bool should_resched(int preempt_offset) in should_resched()
|
| D | refcount.h | 42 static __always_inline void refcount_add(unsigned int i, refcount_t *r) in refcount_add() 51 static __always_inline void refcount_inc(refcount_t *r) in refcount_inc() 59 static __always_inline void refcount_dec(refcount_t *r) in refcount_dec() 67 static __always_inline __must_check 82 static __always_inline __must_check bool refcount_dec_and_test(refcount_t *r) in refcount_dec_and_test() 96 static __always_inline __must_check 121 static __always_inline __must_check bool refcount_inc_not_zero(refcount_t *r) in refcount_inc_not_zero()
|
| /Linux-v5.4/include/linux/unaligned/ |
| D | access_ok.h | 8 static __always_inline u16 get_unaligned_le16(const void *p) in get_unaligned_le16() 13 static __always_inline u32 get_unaligned_le32(const void *p) in get_unaligned_le32() 18 static __always_inline u64 get_unaligned_le64(const void *p) in get_unaligned_le64() 23 static __always_inline u16 get_unaligned_be16(const void *p) in get_unaligned_be16() 28 static __always_inline u32 get_unaligned_be32(const void *p) in get_unaligned_be32() 33 static __always_inline u64 get_unaligned_be64(const void *p) in get_unaligned_be64() 38 static __always_inline void put_unaligned_le16(u16 val, void *p) in put_unaligned_le16() 43 static __always_inline void put_unaligned_le32(u32 val, void *p) in put_unaligned_le32() 48 static __always_inline void put_unaligned_le64(u64 val, void *p) in put_unaligned_le64() 53 static __always_inline void put_unaligned_be16(u16 val, void *p) in put_unaligned_be16() [all …]
|
| /Linux-v5.4/include/asm-generic/ |
| D | preempt.h | 9 static __always_inline int preempt_count(void) in preempt_count() 14 static __always_inline volatile int *preempt_count_ptr(void) in preempt_count_ptr() 19 static __always_inline void preempt_count_set(int pc) in preempt_count_set() 35 static __always_inline void set_preempt_need_resched(void) in set_preempt_need_resched() 39 static __always_inline void clear_preempt_need_resched(void) in clear_preempt_need_resched() 43 static __always_inline bool test_preempt_need_resched(void) in test_preempt_need_resched() 52 static __always_inline void __preempt_count_add(int val) in __preempt_count_add() 57 static __always_inline void __preempt_count_sub(int val) in __preempt_count_sub() 62 static __always_inline bool __preempt_count_dec_and_test(void) in __preempt_count_dec_and_test() 75 static __always_inline bool should_resched(int preempt_offset) in should_resched()
|
| D | qspinlock.h | 20 static __always_inline int queued_spin_is_locked(struct qspinlock *lock) in queued_spin_is_locked() 39 static __always_inline int queued_spin_value_unlocked(struct qspinlock lock) in queued_spin_value_unlocked() 49 static __always_inline int queued_spin_is_contended(struct qspinlock *lock) in queued_spin_is_contended() 58 static __always_inline int queued_spin_trylock(struct qspinlock *lock) in queued_spin_trylock() 74 static __always_inline void queued_spin_lock(struct qspinlock *lock) in queued_spin_lock() 89 static __always_inline void queued_spin_unlock(struct qspinlock *lock) in queued_spin_unlock() 99 static __always_inline bool virt_spin_lock(struct qspinlock *lock) in virt_spin_lock()
|
| /Linux-v5.4/arch/x86/kvm/vmx/ |
| D | ops.h | 22 static __always_inline void vmcs_check16(unsigned long field) in vmcs_check16() 34 static __always_inline void vmcs_check32(unsigned long field) in vmcs_check32() 42 static __always_inline void vmcs_check64(unsigned long field) in vmcs_check64() 54 static __always_inline void vmcs_checkl(unsigned long field) in vmcs_checkl() 66 static __always_inline unsigned long __vmcs_readl(unsigned long field) in __vmcs_readl() 89 static __always_inline u16 vmcs_read16(unsigned long field) in vmcs_read16() 97 static __always_inline u32 vmcs_read32(unsigned long field) in vmcs_read32() 105 static __always_inline u64 vmcs_read64(unsigned long field) in vmcs_read64() 117 static __always_inline unsigned long vmcs_readl(unsigned long field) in vmcs_readl() 155 static __always_inline void __vmcs_writel(unsigned long field, unsigned long value) in __vmcs_writel() [all …]
|
| /Linux-v5.4/arch/x86/include/asm/vdso/ |
| D | gettimeofday.h | 61 static __always_inline 73 static __always_inline 85 static __always_inline 101 static __always_inline 118 static __always_inline 135 static __always_inline 153 static __always_inline long 170 static __always_inline 263 static __always_inline const struct vdso_data *__arch_get_vdso_data(void) in __arch_get_vdso_data() 286 static __always_inline
|
| /Linux-v5.4/arch/mips/include/asm/vdso/ |
| D | gettimeofday.h | 31 static __always_inline long gettimeofday_fallback( in gettimeofday_fallback() 53 static __always_inline long gettimeofday_fallback( in gettimeofday_fallback() 62 static __always_inline long clock_gettime_fallback( in clock_gettime_fallback() 86 static __always_inline int clock_getres_fallback( in clock_getres_fallback() 114 static __always_inline long clock_gettime32_fallback( in clock_gettime32_fallback() 134 static __always_inline int clock_getres32_fallback( in clock_getres32_fallback() 157 static __always_inline u64 read_r4k_count(void) in read_r4k_count() 175 static __always_inline u64 read_gic_count(const struct vdso_data *data) in read_gic_count() 191 static __always_inline u64 __arch_get_hw_counter(s32 clock_mode) in __arch_get_hw_counter() 217 static __always_inline const struct vdso_data *__arch_get_vdso_data(void) in __arch_get_vdso_data()
|
| /Linux-v5.4/arch/riscv/include/asm/ |
| D | atomic.h | 30 static __always_inline int atomic_read(const atomic_t *v) in atomic_read() 34 static __always_inline void atomic_set(atomic_t *v, int i) in atomic_set() 41 static __always_inline s64 atomic64_read(const atomic64_t *v) in atomic64_read() 45 static __always_inline void atomic64_set(atomic64_t *v, s64 i) in atomic64_set() 57 static __always_inline \ 91 static __always_inline \ in ATOMIC_OPS() 103 static __always_inline \ 116 static __always_inline \ 122 static __always_inline \ 202 static __always_inline int atomic_fetch_add_unless(atomic_t *v, int a, int u) [all …]
|
| /Linux-v5.4/include/uapi/linux/byteorder/ |
| D | big_endian.h | 44 static __always_inline __le64 __cpu_to_le64p(const __u64 *p) in __cpu_to_le64p() 48 static __always_inline __u64 __le64_to_cpup(const __le64 *p) in __le64_to_cpup() 52 static __always_inline __le32 __cpu_to_le32p(const __u32 *p) in __cpu_to_le32p() 56 static __always_inline __u32 __le32_to_cpup(const __le32 *p) in __le32_to_cpup() 60 static __always_inline __le16 __cpu_to_le16p(const __u16 *p) in __cpu_to_le16p() 64 static __always_inline __u16 __le16_to_cpup(const __le16 *p) in __le16_to_cpup() 68 static __always_inline __be64 __cpu_to_be64p(const __u64 *p) in __cpu_to_be64p() 72 static __always_inline __u64 __be64_to_cpup(const __be64 *p) in __be64_to_cpup() 76 static __always_inline __be32 __cpu_to_be32p(const __u32 *p) in __cpu_to_be32p() 80 static __always_inline __u32 __be32_to_cpup(const __be32 *p) in __be32_to_cpup() [all …]
|
| D | little_endian.h | 44 static __always_inline __le64 __cpu_to_le64p(const __u64 *p) in __cpu_to_le64p() 48 static __always_inline __u64 __le64_to_cpup(const __le64 *p) in __le64_to_cpup() 52 static __always_inline __le32 __cpu_to_le32p(const __u32 *p) in __cpu_to_le32p() 56 static __always_inline __u32 __le32_to_cpup(const __le32 *p) in __le32_to_cpup() 60 static __always_inline __le16 __cpu_to_le16p(const __u16 *p) in __cpu_to_le16p() 64 static __always_inline __u16 __le16_to_cpup(const __le16 *p) in __le16_to_cpup() 68 static __always_inline __be64 __cpu_to_be64p(const __u64 *p) in __cpu_to_be64p() 72 static __always_inline __u64 __be64_to_cpup(const __be64 *p) in __be64_to_cpup() 76 static __always_inline __be32 __cpu_to_be32p(const __u32 *p) in __cpu_to_be32p() 80 static __always_inline __u32 __be32_to_cpup(const __be32 *p) in __be32_to_cpup() [all …]
|
| /Linux-v5.4/arch/powerpc/include/asm/ |
| D | cmpxchg.h | 88 static __always_inline unsigned long 105 static __always_inline unsigned long 123 static __always_inline unsigned long 140 static __always_inline unsigned long 158 static __always_inline unsigned long 177 static __always_inline unsigned long 222 static __always_inline unsigned long 245 static __always_inline unsigned long 267 static __always_inline unsigned long 295 static __always_inline unsigned long [all …]
|
| /Linux-v5.4/kernel/locking/ |
| D | qspinlock.c | 147 static __always_inline void clear_pending(struct qspinlock *lock) in clear_pending() 160 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) in clear_pending_set_locked() 175 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 193 static __always_inline void clear_pending(struct qspinlock *lock) in clear_pending() 204 static __always_inline void clear_pending_set_locked(struct qspinlock *lock) in clear_pending_set_locked() 219 static __always_inline u32 xchg_tail(struct qspinlock *lock, u32 tail) in xchg_tail() 248 static __always_inline u32 queued_fetch_set_pending_acquire(struct qspinlock *lock) in queued_fetch_set_pending_acquire() 260 static __always_inline void set_locked(struct qspinlock *lock) in set_locked() 271 static __always_inline void __pv_init_node(struct mcs_spinlock *node) { } in __pv_init_node() 272 static __always_inline void __pv_wait_node(struct mcs_spinlock *node, in __pv_wait_node() [all …]
|
| /Linux-v5.4/include/asm-generic/vdso/ |
| D | vsyscall.h | 8 static __always_inline struct vdso_data *__arch_get_k_vdso_data(void) in __arch_get_k_vdso_data() 15 static __always_inline int __arch_update_vdso_data(void) in __arch_update_vdso_data() 22 static __always_inline int __arch_get_clock_mode(struct timekeeper *tk) in __arch_get_clock_mode() 29 static __always_inline void __arch_update_vsyscall(struct vdso_data *vdata, in __arch_update_vsyscall() 36 static __always_inline void __arch_sync_vdso_data(struct vdso_data *vdata) in __arch_sync_vdso_data()
|
| /Linux-v5.4/arch/ia64/include/asm/ |
| D | spinlock.h | 43 static __always_inline void __ticket_spin_lock(arch_spinlock_t *lock) in __ticket_spin_lock() 63 static __always_inline int __ticket_spin_trylock(arch_spinlock_t *lock) in __ticket_spin_trylock() 72 static __always_inline void __ticket_spin_unlock(arch_spinlock_t *lock) in __ticket_spin_unlock() 96 static __always_inline int arch_spin_value_unlocked(arch_spinlock_t lock) in arch_spin_value_unlocked() 112 static __always_inline void arch_spin_lock(arch_spinlock_t *lock) in arch_spin_lock() 117 static __always_inline int arch_spin_trylock(arch_spinlock_t *lock) in arch_spin_trylock() 122 static __always_inline void arch_spin_unlock(arch_spinlock_t *lock) in arch_spin_unlock() 127 static __always_inline void arch_spin_lock_flags(arch_spinlock_t *lock, in arch_spin_lock_flags() 136 static __always_inline void 188 static __always_inline void
|
| /Linux-v5.4/arch/arm64/include/asm/vdso/ |
| D | compat_gettimeofday.h | 21 static __always_inline 39 static __always_inline 56 static __always_inline 73 static __always_inline 94 static __always_inline 115 static __always_inline u64 __arch_get_hw_counter(s32 clock_mode) in __arch_get_hw_counter() 141 static __always_inline const struct vdso_data *__arch_get_vdso_data(void) in __arch_get_vdso_data()
|
| D | gettimeofday.h | 17 static __always_inline 35 static __always_inline 52 static __always_inline 69 static __always_inline u64 __arch_get_hw_counter(s32 clock_mode) in __arch_get_hw_counter() 95 static __always_inline
|
| /Linux-v5.4/include/linux/ |
| D | spinlock.h | 325 static __always_inline raw_spinlock_t *spinlock_check(spinlock_t *lock) in spinlock_check() 336 static __always_inline void spin_lock(spinlock_t *lock) in spin_lock() 341 static __always_inline void spin_lock_bh(spinlock_t *lock) in spin_lock_bh() 346 static __always_inline int spin_trylock(spinlock_t *lock) in spin_trylock() 361 static __always_inline void spin_lock_irq(spinlock_t *lock) in spin_lock_irq() 376 static __always_inline void spin_unlock(spinlock_t *lock) in spin_unlock() 381 static __always_inline void spin_unlock_bh(spinlock_t *lock) in spin_unlock_bh() 386 static __always_inline void spin_unlock_irq(spinlock_t *lock) in spin_unlock_irq() 391 static __always_inline void spin_unlock_irqrestore(spinlock_t *lock, unsigned long flags) in spin_unlock_irqrestore() 396 static __always_inline int spin_trylock_bh(spinlock_t *lock) in spin_trylock_bh() [all …]
|
| D | mm_inline.h | 26 static __always_inline void __update_lru_size(struct lruvec *lruvec, in __update_lru_size() 37 static __always_inline void update_lru_size(struct lruvec *lruvec, in update_lru_size() 47 static __always_inline void add_page_to_lru_list(struct page *page, in add_page_to_lru_list() 54 static __always_inline void add_page_to_lru_list_tail(struct page *page, in add_page_to_lru_list_tail() 61 static __always_inline void del_page_from_lru_list(struct page *page, in del_page_from_lru_list() 90 static __always_inline enum lru_list page_off_lru(struct page *page) in page_off_lru() 114 static __always_inline enum lru_list page_lru(struct page *page) in page_lru()
|
| D | uaccess.h | 58 static __always_inline __must_check unsigned long 66 static __always_inline __must_check unsigned long 88 static __always_inline __must_check unsigned long 96 static __always_inline __must_check unsigned long 140 static __always_inline unsigned long __must_check 148 static __always_inline unsigned long __must_check 156 static __always_inline unsigned long __must_check 166 static __always_inline void pagefault_disabled_inc(void) in pagefault_disabled_inc() 171 static __always_inline void pagefault_disabled_dec(void) in pagefault_disabled_dec() 283 static __always_inline __must_check int
|
| D | rbtree_latch.h | 69 static __always_inline struct latch_tree_node * 75 static __always_inline void 99 static __always_inline void 105 static __always_inline struct latch_tree_node * 143 static __always_inline void 170 static __always_inline void 199 static __always_inline struct latch_tree_node *
|
| D | page-flags.h | 181 static __always_inline int PageTail(struct page *page) in PageTail() 186 static __always_inline int PageCompound(struct page *page) in PageCompound() 247 static __always_inline int Page##uname(struct page *page) \ 251 static __always_inline void SetPage##uname(struct page *page) \ 255 static __always_inline void ClearPage##uname(struct page *page) \ 259 static __always_inline void __SetPage##uname(struct page *page) \ 263 static __always_inline void __ClearPage##uname(struct page *page) \ 267 static __always_inline int TestSetPage##uname(struct page *page) \ 271 static __always_inline int TestClearPage##uname(struct page *page) \ 380 static __always_inline int PageSwapCache(struct page *page) in PAGEFLAG() [all …]
|
| /Linux-v5.4/drivers/misc/mic/scif/ |
| D | scif_map.h | 14 static __always_inline void * 34 static __always_inline void 48 static __always_inline int 69 static __always_inline void 81 static __always_inline void * 95 static __always_inline void 105 static __always_inline int
|