Lines Matching defs:cfs_rq

348 static inline bool list_add_leaf_cfs_rq(struct cfs_rq *cfs_rq)  in list_add_leaf_cfs_rq()
416 static inline void list_del_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_del_leaf_cfs_rq()
442 #define for_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) \ argument
498 static int cfs_rq_is_idle(struct cfs_rq *cfs_rq) in cfs_rq_is_idle()
515 static inline bool list_add_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_add_leaf_cfs_rq()
520 static inline void list_del_leaf_cfs_rq(struct cfs_rq *cfs_rq) in list_del_leaf_cfs_rq()
528 #define for_each_leaf_cfs_rq_safe(rq, cfs_rq, pos) \ argument
546 static int cfs_rq_is_idle(struct cfs_rq *cfs_rq) in cfs_rq_is_idle()
592 static void update_min_vruntime(struct cfs_rq *cfs_rq) in update_min_vruntime()
628 static void __enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in __enqueue_entity()
633 static void __dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in __dequeue_entity()
638 struct sched_entity *__pick_first_entity(struct cfs_rq *cfs_rq) in __pick_first_entity()
659 struct sched_entity *__pick_last_entity(struct cfs_rq *cfs_rq) in __pick_last_entity()
726 static u64 sched_slice(struct cfs_rq *cfs_rq, struct sched_entity *se) in sched_slice()
772 static u64 sched_vslice(struct cfs_rq *cfs_rq, struct sched_entity *se) in sched_vslice()
832 struct cfs_rq *cfs_rq = cfs_rq_of(se); in post_init_entity_util_avg() local
874 static void update_tg_load_avg(struct cfs_rq *cfs_rq) in update_tg_load_avg()
882 static void update_curr(struct cfs_rq *cfs_rq) in update_curr()
928 update_stats_wait_start_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_wait_start_fair()
945 update_stats_wait_end_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_wait_end_fair()
971 update_stats_enqueue_sleeper_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_enqueue_sleeper_fair()
991 update_stats_enqueue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_stats_enqueue_fair()
1008 update_stats_dequeue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_stats_dequeue_fair()
1040 update_stats_curr_start(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_curr_start()
3203 account_entity_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se) in account_entity_enqueue()
3220 account_entity_dequeue(struct cfs_rq *cfs_rq, struct sched_entity *se) in account_entity_dequeue()
3284 enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in enqueue_load_avg()
3291 dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in dequeue_load_avg()
3301 enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { } in enqueue_load_avg()
3303 dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { } in dequeue_load_avg()
3306 static void reweight_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, in reweight_entity()
3336 struct cfs_rq *cfs_rq = cfs_rq_of(se); in reweight_task() local
3421 static long calc_group_shares(struct cfs_rq *cfs_rq) in calc_group_shares()
3489 static inline void cfs_rq_util_change(struct cfs_rq *cfs_rq, int flags) in cfs_rq_util_change()
3536 static inline u64 cfs_rq_last_update_time(struct cfs_rq *cfs_rq) in cfs_rq_last_update_time()
3550 static inline bool child_cfs_rq_on_list(struct cfs_rq *cfs_rq) in child_cfs_rq_on_list()
3568 static inline bool cfs_rq_is_decayed(struct cfs_rq *cfs_rq) in cfs_rq_is_decayed()
3596 static inline void update_tg_load_avg(struct cfs_rq *cfs_rq) in update_tg_load_avg()
3711 update_tg_cfs_util(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_util()
3743 update_tg_cfs_runnable(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_runnable()
3773 update_tg_cfs_load(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_load()
3840 static inline void add_tg_cfs_propagate(struct cfs_rq *cfs_rq, long runnable_sum) in add_tg_cfs_propagate()
3849 struct cfs_rq *cfs_rq, *gcfs_rq; in propagate_entity_load_avg() local
3906 static inline void update_tg_load_avg(struct cfs_rq *cfs_rq) {} in update_tg_load_avg()
3913 static inline void add_tg_cfs_propagate(struct cfs_rq *cfs_rq, long runnable_sum) {} in add_tg_cfs_propagate()
3921 struct cfs_rq *cfs_rq; in migrate_se_pelt_lag() local
4016 update_cfs_rq_load_avg(u64 now, struct cfs_rq *cfs_rq) in update_cfs_rq_load_avg()
4087 static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in attach_entity_load_avg()
4142 static void detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in detach_entity_load_avg()
4173 static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_load_avg()
4221 struct cfs_rq *cfs_rq = cfs_rq_of(se); in sync_entity_load_avg() local
4234 struct cfs_rq *cfs_rq = cfs_rq_of(se); in remove_entity_load_avg() local
4253 static inline unsigned long cfs_rq_runnable_avg(struct cfs_rq *cfs_rq) in cfs_rq_runnable_avg()
4258 static inline unsigned long cfs_rq_load_avg(struct cfs_rq *cfs_rq) in cfs_rq_load_avg()
4296 static inline void util_est_enqueue(struct cfs_rq *cfs_rq, in util_est_enqueue()
4312 static inline void util_est_dequeue(struct cfs_rq *cfs_rq, in util_est_dequeue()
4343 static inline void util_est_update(struct cfs_rq *cfs_rq, in util_est_update()
4459 static inline bool cfs_rq_is_decayed(struct cfs_rq *cfs_rq) in cfs_rq_is_decayed()
4469 static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int not_used1) in update_load_avg()
4477 attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) {} in attach_entity_load_avg()
4479 detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) {} in detach_entity_load_avg()
4487 util_est_enqueue(struct cfs_rq *cfs_rq, struct task_struct *p) {} in util_est_enqueue()
4490 util_est_dequeue(struct cfs_rq *cfs_rq, struct task_struct *p) {} in util_est_dequeue()
4493 util_est_update(struct cfs_rq *cfs_rq, struct task_struct *p, in util_est_update()
4499 static void check_spread(struct cfs_rq *cfs_rq, struct sched_entity *se) in check_spread()
4513 place_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int initial) in place_entity()
4584 enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in enqueue_entity()
4641 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_last() local
4652 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_next() local
4663 struct cfs_rq *cfs_rq = cfs_rq_of(se); in __clear_buddies_skip() local
4671 static void clear_buddies(struct cfs_rq *cfs_rq, struct sched_entity *se) in clear_buddies()
4686 dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in dequeue_entity()
4750 check_preempt_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr) in check_preempt_tick()
4787 set_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in set_next_entity()
4835 pick_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *curr) in pick_next_entity()
4885 static void put_prev_entity(struct cfs_rq *cfs_rq, struct sched_entity *prev) in put_prev_entity()
4910 entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr, int queued) in entity_tick()
5024 struct cfs_rq *cfs_rq, u64 target_runtime) in __assign_cfs_rq_runtime()
5051 static int assign_cfs_rq_runtime(struct cfs_rq *cfs_rq) in assign_cfs_rq_runtime()
5063 static void __account_cfs_rq_runtime(struct cfs_rq *cfs_rq, u64 delta_exec) in __account_cfs_rq_runtime()
5082 void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, u64 delta_exec) in account_cfs_rq_runtime()
5090 static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq) in cfs_rq_throttled()
5096 static inline int throttled_hierarchy(struct cfs_rq *cfs_rq) in throttled_hierarchy()
5121 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_unthrottle_up() local
5139 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_throttle_down() local
5151 static bool throttle_cfs_rq(struct cfs_rq *cfs_rq) in throttle_cfs_rq()
5238 void unthrottle_cfs_rq(struct cfs_rq *cfs_rq) in unthrottle_cfs_rq()
5323 struct cfs_rq *cfs_rq; in distribute_cfs_runtime() local
5474 static void __return_cfs_rq_runtime(struct cfs_rq *cfs_rq) in __return_cfs_rq_runtime()
5497 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) in return_cfs_rq_runtime()
5542 static void check_enqueue_throttle(struct cfs_rq *cfs_rq) in check_enqueue_throttle()
5563 struct cfs_rq *pcfs_rq, *cfs_rq; in sync_throttle() local
5579 static bool check_cfs_rq_runtime(struct cfs_rq *cfs_rq) in check_cfs_rq_runtime()
5680 static void init_cfs_rq_runtime(struct cfs_rq *cfs_rq) in init_cfs_rq_runtime()
5725 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in update_runtime_enabled() local
5743 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in unthrottle_offline_cfs_rqs() local
5772 static void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, u64 delta_exec) {} in account_cfs_rq_runtime()
5773 static bool check_cfs_rq_runtime(struct cfs_rq *cfs_rq) { return false; } in check_cfs_rq_runtime()
5774 static void check_enqueue_throttle(struct cfs_rq *cfs_rq) {} in check_enqueue_throttle()
5776 static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} in return_cfs_rq_runtime()
5778 static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq) in cfs_rq_throttled()
5783 static inline int throttled_hierarchy(struct cfs_rq *cfs_rq) in throttled_hierarchy()
5797 static void init_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} in init_cfs_rq_runtime()
5818 struct cfs_rq *cfs_rq = cfs_rq_of(se); in hrtick_start_fair() local
5891 static bool sched_idle_cfs_rq(struct cfs_rq *cfs_rq) in sched_idle_cfs_rq()
5912 struct cfs_rq *cfs_rq; in enqueue_task_fair() local
6005 struct cfs_rq *cfs_rq; in dequeue_task_fair() local
6112 struct cfs_rq *cfs_rq; in cpu_load_without() local
6135 struct cfs_rq *cfs_rq; in cpu_runnable_without() local
6802 struct cfs_rq *cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util_next() local
7271 struct cfs_rq *cfs_rq = cfs_rq_of(se); in migrate_task_rq_fair() local
7400 struct cfs_rq *cfs_rq = task_cfs_rq(curr); in check_preempt_wakeup() local
7497 struct cfs_rq *cfs_rq; in pick_task_fair() local
7529 struct cfs_rq *cfs_rq = &rq->cfs; in pick_next_task_fair() local
7681 struct cfs_rq *cfs_rq; in put_prev_task_fair() local
7697 struct cfs_rq *cfs_rq = task_cfs_rq(curr); in yield_task_fair() local
8373 static inline bool cfs_rq_has_blocked(struct cfs_rq *cfs_rq) in cfs_rq_has_blocked()
8414 static inline bool cfs_rq_has_blocked(struct cfs_rq *cfs_rq) { return false; } in cfs_rq_has_blocked()
8450 struct cfs_rq *cfs_rq, *pos; in __update_blocked_fair() local
8496 static void update_cfs_rq_h_load(struct cfs_rq *cfs_rq) in update_cfs_rq_h_load()
8531 struct cfs_rq *cfs_rq = task_cfs_rq(p); in task_h_load() local
8540 struct cfs_rq *cfs_rq = &rq->cfs; in __update_blocked_fair() local
11497 struct cfs_rq *cfs_rq = cfs_rq_of(se); in se_fi_update() local
11579 struct cfs_rq *cfs_rq; in task_tick_fair() local
11603 struct cfs_rq *cfs_rq; in task_fork_fair() local
11692 struct cfs_rq *cfs_rq = cfs_rq_of(se); in propagate_entity_cfs_rq() local
11721 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_entity_cfs_rq() local
11743 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_entity_cfs_rq() local
11755 struct cfs_rq *cfs_rq = cfs_rq_of(se); in detach_task_cfs_rq() local
11772 struct cfs_rq *cfs_rq = cfs_rq_of(se); in attach_task_cfs_rq() local
11822 struct cfs_rq *cfs_rq = cfs_rq_of(se); in set_next_task_fair() local
11830 void init_cfs_rq(struct cfs_rq *cfs_rq) in init_cfs_rq()
11877 struct cfs_rq *cfs_rq; in alloc_fair_sched_group() local
11960 void init_tg_cfs_entry(struct task_group *tg, struct cfs_rq *cfs_rq, in init_tg_cfs_entry()
12090 struct cfs_rq *cfs_rq = cfs_rq_of(se); in sched_group_set_idle() local
12199 struct cfs_rq *cfs_rq, *pos; in print_cfs_stats() local