Lines Matching refs:cfs_rq_of
267 static inline struct cfs_rq *cfs_rq_of(struct sched_entity *se) in cfs_rq_of() function
451 static inline struct cfs_rq *cfs_rq_of(struct sched_entity *se) in cfs_rq_of() function
700 cfs_rq = cfs_rq_of(se); in sched_slice()
783 struct cfs_rq *cfs_rq = cfs_rq_of(se); in post_init_entity_util_avg()
870 update_curr(cfs_rq_of(&rq->curr->se)); in update_curr_fair()
2911 struct cfs_rq *cfs_rq = cfs_rq_of(se); in reweight_task()
3100 reweight_entity(cfs_rq_of(se), se, shares, runnable); in update_cfs_group()
3397 cfs_rq = cfs_rq_of(se); in propagate_entity_load_avg()
3650 struct cfs_rq *cfs_rq = cfs_rq_of(se); in sync_entity_load_avg()
3663 struct cfs_rq *cfs_rq = cfs_rq_of(se); in remove_entity_load_avg()
4016 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_last()
4027 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_next()
4038 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_skip()
4508 struct cfs_rq *qcfs_rq = cfs_rq_of(se); in throttle_cfs_rq()
4581 cfs_rq = cfs_rq_of(se); in unthrottle_cfs_rq()
5119 struct cfs_rq *cfs_rq = cfs_rq_of(se); in hrtick_start_fair()
5149 if (cfs_rq_of(&curr->se)->nr_running < sched_nr_latency) in hrtick_update()
5213 cfs_rq = cfs_rq_of(se); in enqueue_task_fair()
5231 cfs_rq = cfs_rq_of(se); in enqueue_task_fair()
5271 cfs_rq = cfs_rq_of(se); in enqueue_task_fair()
5298 cfs_rq = cfs_rq_of(se); in dequeue_task_fair()
5328 cfs_rq = cfs_rq_of(se); in dequeue_task_fair()
6522 struct cfs_rq *cfs_rq = cfs_rq_of(se); in migrate_task_rq_fair()
6641 cfs_rq_of(se)->last = se; in set_last_buddy()
6653 cfs_rq_of(se)->next = se; in set_next_buddy()
6660 cfs_rq_of(se)->skip = se; in set_skip_buddy()
6683 if (unlikely(throttled_hierarchy(cfs_rq_of(pse)))) in check_preempt_wakeup()
6717 update_curr(cfs_rq_of(se)); in check_preempt_wakeup()
6823 put_prev_entity(cfs_rq_of(pse), pse); in pick_next_task_fair()
6827 set_next_entity(cfs_rq_of(se), se); in pick_next_task_fair()
6902 cfs_rq = cfs_rq_of(se); in put_prev_task_fair()
6948 if (!se->on_rq || throttled_hierarchy(cfs_rq_of(se))) in yield_to_task_fair()
7142 (&p->se == cfs_rq_of(&p->se)->next || in task_hot()
7143 &p->se == cfs_rq_of(&p->se)->last)) in task_hot()
7576 update_load_avg(cfs_rq_of(se), se, 0); in update_blocked_averages()
7611 cfs_rq = cfs_rq_of(se); in update_cfs_rq_h_load()
9947 cfs_rq = cfs_rq_of(se); in task_tick_fair()
10057 cfs_rq = cfs_rq_of(se); in propagate_entity_cfs_rq()
10071 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_entity_cfs_rq()
10082 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_entity_cfs_rq()
10102 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_task_cfs_rq()
10119 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_task_cfs_rq()
10169 struct cfs_rq *cfs_rq = cfs_rq_of(se); in set_next_task_fair()
10383 update_load_avg(cfs_rq_of(se), se, UPDATE_TG); in sched_group_set_shares()
10419 rr_interval = NS_TO_JIFFIES(sched_slice(cfs_rq_of(se), se)); in get_rr_interval_fair()