Lines Matching refs:ref

101 int __must_check percpu_ref_init(struct percpu_ref *ref,
104 void percpu_ref_exit(struct percpu_ref *ref);
105 void percpu_ref_switch_to_atomic(struct percpu_ref *ref,
107 void percpu_ref_switch_to_atomic_sync(struct percpu_ref *ref);
108 void percpu_ref_switch_to_percpu(struct percpu_ref *ref);
109 void percpu_ref_kill_and_confirm(struct percpu_ref *ref,
111 void percpu_ref_reinit(struct percpu_ref *ref);
125 static inline void percpu_ref_kill(struct percpu_ref *ref) in percpu_ref_kill() argument
127 percpu_ref_kill_and_confirm(ref, NULL); in percpu_ref_kill()
136 static inline bool __ref_is_percpu(struct percpu_ref *ref, in __ref_is_percpu() argument
152 percpu_ptr = READ_ONCE(ref->percpu_count_ptr); in __ref_is_percpu()
176 static inline void percpu_ref_get_many(struct percpu_ref *ref, unsigned long nr) in percpu_ref_get_many() argument
182 if (__ref_is_percpu(ref, &percpu_count)) in percpu_ref_get_many()
185 atomic_long_add(nr, &ref->count); in percpu_ref_get_many()
198 static inline void percpu_ref_get(struct percpu_ref *ref) in percpu_ref_get() argument
200 percpu_ref_get_many(ref, 1); in percpu_ref_get()
212 static inline bool percpu_ref_tryget(struct percpu_ref *ref) in percpu_ref_tryget() argument
219 if (__ref_is_percpu(ref, &percpu_count)) { in percpu_ref_tryget()
223 ret = atomic_long_inc_not_zero(&ref->count); in percpu_ref_tryget()
246 static inline bool percpu_ref_tryget_live(struct percpu_ref *ref) in percpu_ref_tryget_live() argument
253 if (__ref_is_percpu(ref, &percpu_count)) { in percpu_ref_tryget_live()
256 } else if (!(ref->percpu_count_ptr & __PERCPU_REF_DEAD)) { in percpu_ref_tryget_live()
257 ret = atomic_long_inc_not_zero(&ref->count); in percpu_ref_tryget_live()
275 static inline void percpu_ref_put_many(struct percpu_ref *ref, unsigned long nr) in percpu_ref_put_many() argument
281 if (__ref_is_percpu(ref, &percpu_count)) in percpu_ref_put_many()
283 else if (unlikely(atomic_long_sub_and_test(nr, &ref->count))) in percpu_ref_put_many()
284 ref->release(ref); in percpu_ref_put_many()
298 static inline void percpu_ref_put(struct percpu_ref *ref) in percpu_ref_put() argument
300 percpu_ref_put_many(ref, 1); in percpu_ref_put()
312 static inline bool percpu_ref_is_dying(struct percpu_ref *ref) in percpu_ref_is_dying() argument
314 return ref->percpu_count_ptr & __PERCPU_REF_DEAD; in percpu_ref_is_dying()
325 static inline bool percpu_ref_is_zero(struct percpu_ref *ref) in percpu_ref_is_zero() argument
329 if (__ref_is_percpu(ref, &percpu_count)) in percpu_ref_is_zero()
331 return !atomic_long_read(&ref->count); in percpu_ref_is_zero()