Lines Matching refs:cpu_rq
1496 struct rq *rq = cpu_rq(cpu); in update_numa_stats()
1523 struct rq *rq = cpu_rq(env->dst_cpu); in task_numa_assign()
1534 rq = cpu_rq(env->best_cpu); in task_numa_assign()
1593 struct rq *dst_rq = cpu_rq(env->dst_cpu); in task_numa_compare()
1850 best_rq = cpu_rq(env.best_cpu); in task_numa_migrate()
2280 tsk = READ_ONCE(cpu_rq(cpu)->curr); in task_numa_group()
4876 cfs_rq->throttled_clock_task = rq_clock_task(cpu_rq(cpu)); in sync_throttle()
5367 struct rq *rq = cpu_rq(cpu); in sched_idle_cpu()
5380 return cpu_rq(cpu)->cpu_capacity; in capacity_of()
5385 struct rq *rq = cpu_rq(cpu); in cpu_avg_load_per_task()
5472 if (sync && cpu_rq(this_cpu)->nr_running == 1) in wake_affine_idle()
5485 this_eff_load = cpu_runnable_load(cpu_rq(this_cpu)); in wake_affine_weight()
5503 prev_eff_load = cpu_runnable_load(cpu_rq(prev_cpu)); in wake_affine_weight()
5591 load = cpu_runnable_load(cpu_rq(i)); in find_idlest_group()
5594 avg_load += cfs_rq_load_avg(&cpu_rq(i)->cfs); in find_idlest_group()
5705 struct rq *rq = cpu_rq(i); in find_idlest_group_cpu()
5732 load = cpu_runnable_load(cpu_rq(i)); in find_idlest_group_cpu()
6080 cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util()
6111 cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util_without()
6193 max_cap = cpu_rq(cpu)->rd->max_cpu_capacity; in wake_cap()
6211 struct cfs_rq *cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util_next()
6337 struct root_domain *rd = cpu_rq(smp_processor_id())->rd; in find_energy_efficient_cpu()
7541 struct rq *rq = cpu_rq(cpu); in update_blocked_averages()
7643 struct rq *rq = cpu_rq(cpu); in update_blocked_averages()
7735 struct rq *rq = cpu_rq(cpu); in scale_rt_capacity()
7761 cpu_rq(cpu)->cpu_capacity_orig = arch_scale_cpu_capacity(cpu); in update_cpu_capacity()
7766 cpu_rq(cpu)->cpu_capacity = capacity; in update_cpu_capacity()
7800 struct rq *rq = cpu_rq(cpu); in update_group_capacity()
8024 struct rq *rq = cpu_rq(i); in update_sg_lb_stats()
8609 rq = cpu_rq(i); in find_busiest_queue()
8910 env.dst_rq = cpu_rq(env.new_dst_cpu); in load_balance()
9104 struct rq *target_rq = cpu_rq(target_cpu); in active_load_balance_cpu_stop()
9525 struct rq *rq = cpu_rq(cpu); in nohz_balance_enter_idle()
9631 rq = cpu_rq(balance_cpu); in _nohz_idle_balance()
10289 rq = cpu_rq(i); in online_fair_sched_group()
10316 rq = cpu_rq(cpu); in unregister_fair_sched_group()
10328 struct rq *rq = cpu_rq(cpu); in init_tg_cfs_entry()
10375 struct rq *rq = cpu_rq(i); in sched_group_set_shares()
10478 for_each_leaf_cfs_rq_safe(cpu_rq(cpu), cfs_rq, pos) in print_cfs_stats()