Lines Matching refs:rcu_data
16 static bool rcu_rdp_is_offloaded(struct rcu_data *rdp) in rcu_rdp_is_offloaded()
31 (rdp == this_cpu_ptr(&rcu_data) && in rcu_rdp_is_offloaded()
151 static void rcu_preempt_ctxt_queue(struct rcu_node *rnp, struct rcu_data *rdp) in rcu_preempt_ctxt_queue()
287 if (__this_cpu_read(rcu_data.cpu_no_qs.b.norm)) { in rcu_qs()
289 __this_cpu_read(rcu_data.gp_seq), in rcu_qs()
291 __this_cpu_write(rcu_data.cpu_no_qs.b.norm, false); in rcu_qs()
313 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); in rcu_note_context_switch()
471 struct rcu_data *rdp; in rcu_preempt_deferred_qs_irqrestore()
481 rdp = this_cpu_ptr(&rcu_data); in rcu_preempt_deferred_qs_irqrestore()
586 return (__this_cpu_read(rcu_data.cpu_no_qs.b.exp) || in rcu_preempt_need_deferred_qs()
613 struct rcu_data *rdp; in rcu_preempt_deferred_qs_handler()
615 rdp = container_of(iwp, struct rcu_data, defer_qs_iw); in rcu_preempt_deferred_qs_handler()
639 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); in rcu_read_unlock_special()
739 __this_cpu_read(rcu_data.core_needs_qs) && in rcu_flavor_sched_clock_irq()
740 __this_cpu_read(rcu_data.cpu_no_qs.b.norm) && in rcu_flavor_sched_clock_irq()
781 struct rcu_data *rdp; in dump_blkd_tasks()
803 rdp = per_cpu_ptr(&rcu_data, cpu); in dump_blkd_tasks()
820 struct rcu_data *rdp; in rcu_read_unlock_strict()
824 rdp = this_cpu_ptr(&rcu_data); in rcu_read_unlock_strict()
849 if (!__this_cpu_read(rcu_data.cpu_no_qs.s)) in rcu_qs()
852 __this_cpu_read(rcu_data.gp_seq), TPS("cpuqs")); in rcu_qs()
853 __this_cpu_write(rcu_data.cpu_no_qs.b.norm, false); in rcu_qs()
854 if (__this_cpu_read(rcu_data.cpu_no_qs.b.exp)) in rcu_qs()
855 rcu_report_exp_rdp(this_cpu_ptr(&rcu_data)); in rcu_qs()
869 if (!raw_cpu_read(rcu_data.rcu_urgent_qs)) in rcu_all_qs()
873 if (!smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs))) { in rcu_all_qs()
877 this_cpu_write(rcu_data.rcu_urgent_qs, false); in rcu_all_qs()
878 if (unlikely(raw_cpu_read(rcu_data.rcu_need_heavy_qs))) { in rcu_all_qs()
896 if (!smp_load_acquire(this_cpu_ptr(&rcu_data.rcu_urgent_qs))) in rcu_note_context_switch()
898 this_cpu_write(rcu_data.rcu_urgent_qs, false); in rcu_note_context_switch()
899 if (unlikely(raw_cpu_read(rcu_data.rcu_need_heavy_qs))) in rcu_note_context_switch()
942 struct rcu_data *rdp = this_cpu_ptr(&rcu_data); in rcu_preempt_deferred_qs()
1005 struct rcu_data *rdp = per_cpu_ptr(&rcu_data, cpu); in rcu_cpu_kthread_setup()
1016 static bool rcu_is_callbacks_nocb_kthread(struct rcu_data *rdp) in rcu_is_callbacks_nocb_kthread()
1029 static bool rcu_is_callbacks_kthread(struct rcu_data *rdp) in rcu_is_callbacks_kthread()