Lines Matching refs:rq_of

253 static inline struct rq *rq_of(struct cfs_rq *cfs_rq)  in rq_of()  function
288 struct rq *rq = rq_of(cfs_rq); in list_add_leaf_cfs_rq()
414 static inline struct rq *rq_of(struct cfs_rq *cfs_rq) in rq_of() function
751 long cpu_scale = arch_scale_cpu_capacity(NULL, cpu_of(rq_of(cfs_rq))); in post_init_entity_util_avg()
805 u64 now = rq_clock_task(rq_of(cfs_rq)); in update_curr()
850 wait_start = rq_clock(rq_of(cfs_rq)); in update_stats_wait_start()
869 delta = rq_clock(rq_of(cfs_rq)) - schedstat_val(se->statistics.wait_start); in update_stats_wait_end()
908 u64 delta = rq_clock(rq_of(cfs_rq)) - sleep_start; in update_stats_enqueue_sleeper()
925 u64 delta = rq_clock(rq_of(cfs_rq)) - block_start; in update_stats_enqueue_sleeper()
999 rq_clock(rq_of(cfs_rq))); in update_stats_dequeue()
1002 rq_clock(rq_of(cfs_rq))); in update_stats_dequeue()
1015 se->exec_start = rq_clock_task(rq_of(cfs_rq)); in update_stats_curr_start()
2695 update_load_add(&rq_of(cfs_rq)->load, se->load.weight); in account_entity_enqueue()
2698 struct rq *rq = rq_of(cfs_rq); in account_entity_enqueue()
2712 update_load_sub(&rq_of(cfs_rq)->load, se->load.weight); in account_entity_dequeue()
2715 account_numa_dequeue(rq_of(cfs_rq), task_of(se)); in account_entity_dequeue()
3038 struct rq *rq = rq_of(cfs_rq); in cfs_rq_util_change()
3137 __update_load_avg_blocked_se(p_last_update_time, cpu_of(rq_of(prev)), se); in set_task_rq_fair()
3276 arch_scale_cpu_capacity(NULL, cpu_of(rq_of(cfs_rq))); in update_tg_cfs_runnable()
3424 decayed |= __update_load_avg_cfs_rq(now, cpu_of(rq_of(cfs_rq)), cfs_rq); in update_cfs_rq_load_avg()
3515 struct rq *rq = rq_of(cfs_rq); in update_load_avg()
3576 __update_load_avg_blocked_se(last_update_time, cpu_of(rq_of(cfs_rq)), se); in sync_entity_load_avg()
4021 resched_curr(rq_of(cfs_rq)); in check_preempt_tick()
4045 resched_curr(rq_of(cfs_rq)); in check_preempt_tick()
4071 if (schedstat_enabled() && rq_of(cfs_rq)->load.weight >= 2*se->load.weight) { in set_next_entity()
4187 resched_curr(rq_of(cfs_rq)); in entity_tick()
4194 hrtimer_active(&rq_of(cfs_rq)->hrtick_timer)) in entity_tick()
4281 return rq_clock_task(rq_of(cfs_rq)) - cfs_rq->throttled_clock_task_time; in cfs_rq_clock_task()
4334 if (likely((s64)(rq_clock(rq_of(cfs_rq)) - cfs_rq->runtime_expires) < 0)) in expire_cfs_rq_runtime()
4371 resched_curr(rq_of(cfs_rq)); in __account_cfs_rq_runtime()
4441 struct rq *rq = rq_of(cfs_rq); in throttle_cfs_rq()
4447 se = cfs_rq->tg->se[cpu_of(rq_of(cfs_rq))]; in throttle_cfs_rq()
4499 struct rq *rq = rq_of(cfs_rq); in unthrottle_cfs_rq()
4554 struct rq *rq = rq_of(cfs_rq); in distribute_cfs_runtime()
4980 return rq_clock_task(rq_of(cfs_rq)); in cfs_rq_clock_task()
7422 struct rq *rq = rq_of(cfs_rq); in update_cfs_rq_h_load()