Lines Matching refs:cpu_of

304 	int cpu = cpu_of(rq);  in list_add_leaf_cfs_rq()
792 long cpu_scale = arch_scale_cpu_capacity(cpu_of(rq_of(cfs_rq))); in post_init_entity_util_avg()
4004 cpu = cpu_of(rq_of(cfs_rq)); in util_est_dequeue()
4049 if (task_fits_capacity(p, capacity_of(cpu_of(rq)))) { in update_misfit_status()
4711 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_unthrottle_up()
4729 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_throttle_down()
4769 se = cfs_rq->tg->se[cpu_of(rq_of(cfs_rq))]; in throttle_cfs_rq()
4817 se = cfs_rq->tg->se[cpu_of(rq)]; in unthrottle_cfs_rq()
5299 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in update_runtime_enabled()
5317 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in unthrottle_offline_cfs_rqs()
5688 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_load_without()
5711 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_runnable_without()
6019 int core = cpu_of(rq); in __update_idle_core()
7874 thermal_pressure = arch_scale_thermal_pressure(cpu_of(rq)); in __update_blocked_others()
7910 int cpu = cpu_of(rq); in __update_blocked_fair()
7954 struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)]; in update_cfs_rq_h_load()
9398 util = cpu_util(cpu_of(rq)); in find_busiest_queue()
9676 __cpumask_clear_cpu(cpu_of(busiest), cpus); in load_balance()
9735 stop_one_cpu_nowait(cpu_of(busiest), in load_balance()
9852 int busiest_cpu = cpu_of(busiest_rq); in active_load_balance_cpu_stop()
11307 return cfs_rq ? cpu_of(rq_of(cfs_rq)) : -1; in sched_trace_cfs_rq_cpu()
11343 return rq ? cpu_of(rq) : -1; in sched_trace_rq_cpu()