Lines Matching refs:kfree_rcu_cpu
3069 struct kfree_rcu_cpu *krcp;
3094 struct kfree_rcu_cpu { struct
3107 static DEFINE_PER_CPU(struct kfree_rcu_cpu, krc) = { argument
3122 static inline struct kfree_rcu_cpu *
3125 struct kfree_rcu_cpu *krcp; in krc_this_cpu_lock()
3135 krc_this_cpu_unlock(struct kfree_rcu_cpu *krcp, unsigned long flags) in krc_this_cpu_unlock()
3142 get_cached_bnode(struct kfree_rcu_cpu *krcp) in get_cached_bnode()
3153 put_cached_bnode(struct kfree_rcu_cpu *krcp, in put_cached_bnode()
3175 struct kfree_rcu_cpu *krcp; in kfree_rcu_work()
3260 static inline bool queue_kfree_rcu_work(struct kfree_rcu_cpu *krcp) in queue_kfree_rcu_work()
3316 static inline void kfree_rcu_drain_unlock(struct kfree_rcu_cpu *krcp, in kfree_rcu_drain_unlock()
3340 struct kfree_rcu_cpu *krcp = container_of(work, struct kfree_rcu_cpu, in kfree_rcu_monitor()
3351 kvfree_call_rcu_add_ptr_to_bulk(struct kfree_rcu_cpu *krcp, void *ptr) in kvfree_call_rcu_add_ptr_to_bulk()
3424 struct kfree_rcu_cpu *krcp; in kvfree_call_rcu()
3504 struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu); in kfree_rcu_shrink_count()
3520 struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu); in kfree_rcu_shrink_scan()
3552 struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu); in kfree_rcu_scheduler_running()
4451 struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu); in kfree_rcu_batch_init()