Searched refs:this_rq (Results 1 – 8 of 8) sorted by relevance
| /Linux-v5.10/kernel/sched/ |
| D | loadavg.c | 79 long calc_load_fold_active(struct rq *this_rq, long adjust) in calc_load_fold_active() argument 83 nr_active = this_rq->nr_running - adjust; in calc_load_fold_active() 84 nr_active += (long)this_rq->nr_uninterruptible; in calc_load_fold_active() 86 if (nr_active != this_rq->calc_load_active) { in calc_load_fold_active() 87 delta = nr_active - this_rq->calc_load_active; in calc_load_fold_active() 88 this_rq->calc_load_active = nr_active; in calc_load_fold_active() 252 calc_load_nohz_fold(this_rq()); in calc_load_nohz_start() 266 struct rq *this_rq = this_rq(); in calc_load_nohz_stop() local 271 this_rq->calc_load_update = READ_ONCE(calc_load_update); in calc_load_nohz_stop() 272 if (time_before(jiffies, this_rq->calc_load_update)) in calc_load_nohz_stop() [all …]
|
| D | sched.h | 99 extern void calc_global_load_tick(struct rq *this_rq); 100 extern long calc_load_fold_active(struct rq *this_rq, long adjust); 1095 #define this_rq() this_cpu_ptr(&runqueues) macro 1335 rq = this_rq(); in this_rq_lock_irq() 1802 void (*task_woken)(struct rq *this_rq, struct task_struct *task); 1820 void (*switched_from)(struct rq *this_rq, struct task_struct *task); 1821 void (*switched_to) (struct rq *this_rq, struct task_struct *task); 1822 void (*prio_changed) (struct rq *this_rq, struct task_struct *task, 2089 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument 2090 __releases(this_rq->lock) in _double_lock_balance() [all …]
|
| D | rt.c | 263 static void pull_rt_task(struct rq *this_rq); 425 static inline void pull_rt_task(struct rq *this_rq) in pull_rt_task() argument 584 return this_rq()->rd->span; in sched_rt_period_mask() 2090 rq = this_rq(); in rto_push_irq_work_func() 2119 static void pull_rt_task(struct rq *this_rq) in pull_rt_task() argument 2121 int this_cpu = this_rq->cpu, cpu; in pull_rt_task() 2125 int rt_overload_count = rt_overloaded(this_rq); in pull_rt_task() 2138 cpumask_test_cpu(this_rq->cpu, this_rq->rd->rto_mask)) in pull_rt_task() 2143 tell_cpu_to_push(this_rq); in pull_rt_task() 2148 for_each_cpu(cpu, this_rq->rd->rto_mask) { in pull_rt_task() [all …]
|
| D | fair.c | 3886 static int newidle_balance(struct rq *this_rq, struct rq_flags *rf); 6137 avg_idle = this_rq()->avg_idle / 512; in select_idle_cpu() 6250 this_rq()->nr_running <= 1) { in select_idle_sibling() 9532 static int load_balance(int this_cpu, struct rq *this_rq, in load_balance() argument 9546 .dst_rq = this_rq, in load_balance() 10247 SCHED_WARN_ON(rq != this_rq()); in nohz_balance_exit_idle() 10342 static bool _nohz_idle_balance(struct rq *this_rq, unsigned int flags, in _nohz_idle_balance() argument 10350 int this_cpu = this_rq->cpu; in _nohz_idle_balance() 10423 has_blocked_load |= this_rq->has_blocked_load; in _nohz_idle_balance() 10427 rebalance_domains(this_rq, CPU_IDLE); in _nohz_idle_balance() [all …]
|
| D | cputime.c | 221 struct rq *rq = this_rq(); in account_idle_time() 241 steal -= this_rq()->prev_steal_time; in steal_account_process_time() 244 this_rq()->prev_steal_time += steal; in steal_account_process_time() 384 } else if (p == this_rq()->idle) { in irqtime_account_process_tick() 493 else if ((p != this_rq()->idle) || (irq_count() != HARDIRQ_OFFSET)) in account_process_tick()
|
| D | deadline.c | 2212 static void pull_dl_task(struct rq *this_rq) in pull_dl_task() argument 2214 int this_cpu = this_rq->cpu, cpu; in pull_dl_task() 2220 if (likely(!dl_overloaded(this_rq))) in pull_dl_task() 2229 for_each_cpu(cpu, this_rq->rd->dlo_mask) { in pull_dl_task() 2239 if (this_rq->dl.dl_nr_running && in pull_dl_task() 2240 dl_time_before(this_rq->dl.earliest_dl.curr, in pull_dl_task() 2245 double_lock_balance(this_rq, src_rq); in pull_dl_task() 2262 (!this_rq->dl.dl_nr_running || in pull_dl_task() 2264 this_rq->dl.earliest_dl.curr))) { in pull_dl_task() 2280 activate_task(this_rq, p, 0); in pull_dl_task() [all …]
|
| D | idle.c | 22 idle_set_state(this_rq(), idle_state); in sched_idle_set_state()
|
| D | core.c | 403 if (rq == this_rq()) in hrtick_start() 1788 struct rq *rq = this_rq(); in migration_cpu_stop() 2424 rq = this_rq(); in ttwu_stat() 2562 struct rq *rq = this_rq(); in sched_ttwu_pending() 3581 struct rq *rq = this_rq(); in finish_task_switch() 4302 schedstat_inc(this_rq()->sched_count); in schedule_debug() 6213 rq = this_rq(); in yield_to() 6687 BUG_ON(current != this_rq()->idle); in idle_task_exit()
|