Lines Matching refs:cpu_rq

2053 		struct rq *rq = cpu_rq(cpu);  in update_numa_stats()
2085 struct rq *rq = cpu_rq(env->dst_cpu); in task_numa_assign()
2100 rq = cpu_rq(env->dst_cpu); in task_numa_assign()
2115 rq = cpu_rq(env->best_cpu); in task_numa_assign()
2174 struct rq *dst_rq = cpu_rq(env->dst_cpu); in task_numa_compare()
2533 best_rq = cpu_rq(env.best_cpu); in task_numa_migrate()
2976 tsk = READ_ONCE(cpu_rq(cpu)->curr); in task_numa_group()
5872 rq = cpu_rq(this_cpu); in distribute_cfs_runtime()
6093 cfs_rq->throttled_clock_pelt = rq_clock_pelt(cpu_rq(cpu)); in sync_throttle()
6247 struct rq *rq = cpu_rq(i); in destroy_cfs_bandwidth()
6472 unsigned long rq_util_min = uclamp_rq_get(cpu_rq(cpu), UCLAMP_MIN); in cpu_overutilized()
6473 unsigned long rq_util_max = uclamp_rq_get(cpu_rq(cpu), UCLAMP_MAX); in cpu_overutilized()
6500 return sched_idle_rq(cpu_rq(cpu)); in sched_idle_cpu()
6754 return cpu_rq(cpu)->cpu_capacity; in capacity_of()
6834 if (sync && cpu_rq(this_cpu)->nr_running == 1) in wake_affine_idle()
6850 this_eff_load = cpu_load(cpu_rq(this_cpu)); in wake_affine_weight()
6868 prev_eff_load = cpu_load(cpu_rq(prev_cpu)); in wake_affine_weight()
6928 struct rq *rq = cpu_rq(i); in find_idlest_group_cpu()
6958 load = cpu_load(cpu_rq(i)); in find_idlest_group_cpu()
7025 sched_cpu_cookie_match(cpu_rq(cpu), p)) in __select_idle_cpu()
7286 cpu_cap = capacity_orig_of(cpu) - thermal_load_avg(cpu_rq(cpu)); in select_idle_capacity()
7469 struct cfs_rq *cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util()
7589 unsigned long irq = cpu_util_irq(cpu_rq(prev_cpu)); in eenv_task_busy_time()
7788 struct rq *rq = cpu_rq(cpu); in find_energy_efficient_cpu()
8619 if (!sched_core_cookie_match(cpu_rq(env->dst_cpu), p)) in task_hot()
9202 struct rq *rq = cpu_rq(cpu); in update_blocked_averages()
9282 struct rq *rq = cpu_rq(cpu); in scale_rt_capacity()
9315 cpu_rq(cpu)->cpu_capacity_orig = arch_scale_cpu_capacity(cpu); in update_cpu_capacity()
9320 cpu_rq(cpu)->cpu_capacity = capacity; in update_cpu_capacity()
9321 trace_sched_cpu_capacity_tp(cpu_rq(cpu)); in update_cpu_capacity()
9680 struct rq *rq = cpu_rq(i); in update_sg_lb_stats()
9965 struct rq *rq = cpu_rq(cpu); in idle_cpu_without()
10005 struct rq *rq = cpu_rq(i); in update_sg_wakeup_stats()
10126 if (!sched_group_cookie_match(cpu_rq(this_cpu), p, group)) in find_idlest_group()
10783 rq = cpu_rq(i); in find_busiest_queue()
11172 env.dst_rq = cpu_rq(env.new_dst_cpu); in load_balance()
11365 struct rq *target_rq = cpu_rq(target_cpu); in active_load_balance_cpu_stop()
11628 smp_call_function_single_async(ilb_cpu, &cpu_rq(ilb_cpu)->nohz_csd); in kick_ilb()
11805 struct rq *rq = cpu_rq(cpu); in nohz_balance_enter_idle()
11938 rq = cpu_rq(balance_cpu); in _nohz_idle_balance()
12018 _nohz_idle_balance(cpu_rq(cpu), NOHZ_STATS_KICK); in nohz_run_idle_balance()
12370 cfs_rq = &cpu_rq(cpu)->cfs; in task_is_throttled_fair()
12677 rq = cpu_rq(i); in online_fair_sched_group()
12706 rq = cpu_rq(cpu); in unregister_fair_sched_group()
12718 struct rq *rq = cpu_rq(cpu); in init_tg_cfs_entry()
12766 struct rq *rq = cpu_rq(i); in __sched_group_set_shares()
12817 struct rq *rq = cpu_rq(i); in sched_group_set_idle()
12960 for_each_leaf_cfs_rq_safe(cpu_rq(cpu), cfs_rq, pos) in print_cfs_stats()
13002 INIT_CSD(&cpu_rq(i)->cfsb_csd, __cfsb_csd_unthrottle, cpu_rq(i)); in init_sched_fair_class()
13003 INIT_LIST_HEAD(&cpu_rq(i)->cfsb_csd_list); in init_sched_fair_class()