Lines Matching refs:cpu_of
334 int cpu = cpu_of(rq); in list_add_leaf_cfs_rq()
1102 long cpu_scale = arch_scale_cpu_capacity(cpu_of(rq_of(cfs_rq))); in post_init_entity_util_avg()
4339 now += sched_clock_cpu(cpu_of(rq)) - u64_u32_load(rq->clock_idle); in migrate_se_pelt_lag()
4748 if (task_util(p) > capacity_orig_of(cpu_of(rq_of(cfs_rq)))) in util_est_update()
4919 if (task_fits_cpu(p, cpu_of(rq))) { in update_misfit_status()
5515 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_unthrottle_up()
5544 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_throttle_down()
5588 se = cfs_rq->tg->se[cpu_of(rq_of(cfs_rq))]; in throttle_cfs_rq()
5656 se = cfs_rq->tg->se[cpu_of(rq)]; in unthrottle_cfs_rq()
5792 smp_call_function_single_async(cpu_of(rq), &rq->cfsb_csd); in __unthrottle_cfs_rq_async()
5857 if (cpu_of(rq) != this_cpu || in distribute_cfs_runtime()
6277 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in update_runtime_enabled()
6302 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in unthrottle_offline_cfs_rqs()
6344 int cpu = cpu_of(rq); in sched_fair_update_stop_tick()
6717 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_load_without()
6740 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_runnable_without()
7064 int core = cpu_of(rq); in __update_idle_core()
9076 thermal_pressure = arch_scale_thermal_pressure(cpu_of(rq)); in __update_blocked_others()
9095 int cpu = cpu_of(rq); in __update_blocked_fair()
9142 struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)]; in update_cfs_rq_h_load()
11197 __cpumask_clear_cpu(cpu_of(busiest), cpus); in load_balance()
11257 stop_one_cpu_nowait(cpu_of(busiest), in load_balance()
11363 int busiest_cpu = cpu_of(busiest_rq); in active_load_balance_cpu_stop()
12225 if (unlikely(on_null_domain(rq) || !cpu_active(cpu_of(rq)))) in trigger_load_balance()