Lines Matching refs:cfs_rq_of
274 static inline struct cfs_rq *cfs_rq_of(struct sched_entity *se) in cfs_rq_of() function
428 static inline struct cfs_rq *cfs_rq_of(struct sched_entity *se) in cfs_rq_of() function
666 cfs_rq = cfs_rq_of(se); in sched_slice()
749 struct cfs_rq *cfs_rq = cfs_rq_of(se); in post_init_entity_util_avg()
839 update_curr(cfs_rq_of(&rq->curr->se)); in update_curr_fair()
2838 struct cfs_rq *cfs_rq = cfs_rq_of(se); in reweight_task()
3027 reweight_entity(cfs_rq_of(se), se, shares, runnable); in update_cfs_group()
3324 cfs_rq = cfs_rq_of(se); in propagate_entity_load_avg()
3572 struct cfs_rq *cfs_rq = cfs_rq_of(se); in sync_entity_load_avg()
3585 struct cfs_rq *cfs_rq = cfs_rq_of(se); in remove_entity_load_avg()
3912 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_last()
3923 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_next()
3934 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_skip()
4456 struct cfs_rq *qcfs_rq = cfs_rq_of(se); in throttle_cfs_rq()
4527 cfs_rq = cfs_rq_of(se); in unthrottle_cfs_rq()
5029 struct cfs_rq *cfs_rq = cfs_rq_of(se); in hrtick_start_fair()
5059 if (cfs_rq_of(&curr->se)->nr_running < sched_nr_latency) in hrtick_update()
5103 cfs_rq = cfs_rq_of(se); in enqueue_task_fair()
5120 cfs_rq = cfs_rq_of(se); in enqueue_task_fair()
5150 cfs_rq = cfs_rq_of(se); in dequeue_task_fair()
5179 cfs_rq = cfs_rq_of(se); in dequeue_task_fair()
6380 struct cfs_rq *cfs_rq = cfs_rq_of(se); in migrate_task_rq_fair()
6490 cfs_rq_of(se)->last = se; in set_last_buddy()
6502 cfs_rq_of(se)->next = se; in set_next_buddy()
6509 cfs_rq_of(se)->skip = se; in set_skip_buddy()
6532 if (unlikely(throttled_hierarchy(cfs_rq_of(pse)))) in check_preempt_wakeup()
6566 update_curr(cfs_rq_of(se)); in check_preempt_wakeup()
6672 put_prev_entity(cfs_rq_of(pse), pse); in pick_next_task_fair()
6676 set_next_entity(cfs_rq_of(se), se); in pick_next_task_fair()
6740 cfs_rq = cfs_rq_of(se); in put_prev_task_fair()
6786 if (!se->on_rq || throttled_hierarchy(cfs_rq_of(se))) in yield_to_task_fair()
6972 (&p->se == cfs_rq_of(&p->se)->next || in task_hot()
6973 &p->se == cfs_rq_of(&p->se)->last)) in task_hot()
7385 update_load_avg(cfs_rq_of(se), se, 0); in update_blocked_averages()
7432 cfs_rq = cfs_rq_of(se); in update_cfs_rq_h_load()
9655 cfs_rq = cfs_rq_of(se); in task_tick_fair()
9762 cfs_rq = cfs_rq_of(se); in propagate_entity_cfs_rq()
9776 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_entity_cfs_rq()
9787 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_entity_cfs_rq()
9807 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_task_cfs_rq()
9824 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_task_cfs_rq()
9864 struct cfs_rq *cfs_rq = cfs_rq_of(se); in set_curr_task_fair()
10078 update_load_avg(cfs_rq_of(se), se, UPDATE_TG); in sched_group_set_shares()
10114 rr_interval = NS_TO_JIFFIES(sched_slice(cfs_rq_of(se), se)); in get_rr_interval_fair()