Lines Matching refs:kfree_rcu_cpu
2812 struct kfree_rcu_cpu *krcp;
2842 struct kfree_rcu_cpu { struct
2867 static DEFINE_PER_CPU(struct kfree_rcu_cpu, krc) = { argument
2882 static inline struct kfree_rcu_cpu *
2885 struct kfree_rcu_cpu *krcp; in krc_this_cpu_lock()
2895 krc_this_cpu_unlock(struct kfree_rcu_cpu *krcp, unsigned long flags) in krc_this_cpu_unlock()
2901 get_cached_bnode(struct kfree_rcu_cpu *krcp) in get_cached_bnode()
2912 put_cached_bnode(struct kfree_rcu_cpu *krcp, in put_cached_bnode()
2925 drain_page_cache(struct kfree_rcu_cpu *krcp) in drain_page_cache()
2948 kvfree_rcu_bulk(struct kfree_rcu_cpu *krcp, in kvfree_rcu_bulk()
3017 struct kfree_rcu_cpu *krcp; in kfree_rcu_work()
3056 need_offload_krc(struct kfree_rcu_cpu *krcp) in need_offload_krc()
3079 static int krc_count(struct kfree_rcu_cpu *krcp) in krc_count()
3091 schedule_delayed_monitor_work(struct kfree_rcu_cpu *krcp) in schedule_delayed_monitor_work()
3106 kvfree_rcu_drain_ready(struct kfree_rcu_cpu *krcp) in kvfree_rcu_drain_ready()
3148 struct kfree_rcu_cpu *krcp = container_of(work, in kfree_rcu_monitor()
3149 struct kfree_rcu_cpu, monitor_work.work); in kfree_rcu_monitor()
3212 struct kfree_rcu_cpu *krcp = in schedule_page_work_fn()
3213 container_of(t, struct kfree_rcu_cpu, hrtimer); in schedule_page_work_fn()
3222 struct kfree_rcu_cpu *krcp = in fill_page_cache_func()
3223 container_of(work, struct kfree_rcu_cpu, in fill_page_cache_func()
3255 run_page_cache_worker(struct kfree_rcu_cpu *krcp) in run_page_cache_worker()
3282 add_ptr_to_bulk_krc_lock(struct kfree_rcu_cpu **krcp, in add_ptr_to_bulk_krc_lock()
3349 struct kfree_rcu_cpu *krcp; in kvfree_call_rcu()
3419 struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu); in kfree_rcu_shrink_count()
3436 struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu); in kfree_rcu_shrink_scan()
3464 struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu); in kfree_rcu_scheduler_running()
4948 struct kfree_rcu_cpu *krcp = per_cpu_ptr(&krc, cpu); in kfree_rcu_batch_init()