Lines Matching refs:cfs

416 	return container_of(cfs_rq, struct rq, cfs);  in rq_of()
425 return &task_rq(p)->cfs; in task_cfs_rq()
433 return &rq->cfs; in cfs_rq_of()
451 for (cfs_rq = &rq->cfs, pos = NULL; cfs_rq; cfs_rq = pos)
3040 if (&rq->cfs == cfs_rq || (flags & SCHED_CPUFREQ_MIGRATION)) { in cfs_rq_util_change()
4540 if (rq->curr == rq->idle && rq->cfs.nr_running) in unthrottle_cfs_rq()
5033 if (rq->cfs.h_nr_running > 1) { in hrtick_start_fair()
5090 util_est_enqueue(&rq->cfs, p); in enqueue_task_fair()
5192 util_est_dequeue(&rq->cfs, p, task_sleep); in dequeue_task_fair()
5358 return cfs_rq_runnable_load_avg(&rq->cfs); in weighted_cpuload()
5520 unsigned long nr_running = READ_ONCE(rq->cfs.h_nr_running); in cpu_avg_load_per_task()
5737 avg_load += cfs_rq_load_avg(&cpu_rq(i)->cfs); in find_idlest_group()
6207 cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util()
6229 cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util_wake()
6601 struct cfs_rq *cfs_rq = &rq->cfs; in pick_next_task_fair()
6644 cfs_rq = &rq->cfs; in pick_next_task_fair()
7465 struct cfs_rq *cfs_rq = &rq->cfs; in update_blocked_averages()
7864 sgs->sum_nr_running += rq->cfs.h_nr_running; in update_sg_lb_stats()
8482 (env->src_rq->cfs.h_nr_running == 1)) { in need_active_balance()
9152 if ((rq->cfs.h_nr_running >= 1) && in nohz_balancer_kick()
9564 if (this_rq->cfs.h_nr_running && !pulled_task) in idle_balance()
9572 if (this_rq->nr_running != this_rq->cfs.h_nr_running) in idle_balance()
10037 se->cfs_rq = &rq->cfs; in init_tg_cfs_entry()
10113 if (rq->cfs.load.weight) in get_rr_interval_fair()