Lines Matching refs:avg
787 struct sched_avg *sa = &se->avg; in init_entity_runnable_average()
833 struct sched_avg *sa = &se->avg; in post_init_entity_util_avg()
835 long cap = (long)(cpu_scale - cfs_rq->avg.util_avg) / 2; in post_init_entity_util_avg()
848 se->avg.last_update_time = cfs_rq_clock_pelt(cfs_rq); in post_init_entity_util_avg()
853 if (cfs_rq->avg.util_avg != 0) { in post_init_entity_util_avg()
854 sa->util_avg = cfs_rq->avg.util_avg * se->load.weight; in post_init_entity_util_avg()
855 sa->util_avg /= (cfs_rq->avg.load_avg + 1); in post_init_entity_util_avg()
2442 delta = p->se.avg.load_sum; in numa_get_avg_runtime()
3286 cfs_rq->avg.load_avg += se->avg.load_avg; in enqueue_load_avg()
3287 cfs_rq->avg.load_sum += se_weight(se) * se->avg.load_sum; in enqueue_load_avg()
3293 sub_positive(&cfs_rq->avg.load_avg, se->avg.load_avg); in dequeue_load_avg()
3294 sub_positive(&cfs_rq->avg.load_sum, se_weight(se) * se->avg.load_sum); in dequeue_load_avg()
3296 cfs_rq->avg.load_sum = max_t(u32, cfs_rq->avg.load_sum, in dequeue_load_avg()
3297 cfs_rq->avg.load_avg * PELT_MIN_DIVIDER); in dequeue_load_avg()
3321 u32 divider = get_pelt_divider(&se->avg); in reweight_entity()
3323 se->avg.load_avg = div_u64(se_weight(se) * se->avg.load_sum, divider); in reweight_entity()
3428 load = max(scale_load_down(cfs_rq->load.weight), cfs_rq->avg.load_avg); in calc_group_shares()
3538 return u64_u32_load_copy(cfs_rq->avg.last_update_time, in cfs_rq_last_update_time()
3573 if (!load_avg_is_decayed(&cfs_rq->avg)) in cfs_rq_is_decayed()
3598 long delta = cfs_rq->avg.load_avg - cfs_rq->tg_load_avg_contrib; in update_tg_load_avg()
3608 cfs_rq->tg_load_avg_contrib = cfs_rq->avg.load_avg; in update_tg_load_avg()
3633 if (!(se->avg.last_update_time && prev)) in set_task_rq_fair()
3640 se->avg.last_update_time = n_last_update_time; in set_task_rq_fair()
3713 long delta_sum, delta_avg = gcfs_rq->avg.util_avg - se->avg.util_avg; in update_tg_cfs_util()
3724 divider = get_pelt_divider(&cfs_rq->avg); in update_tg_cfs_util()
3728 se->avg.util_avg = gcfs_rq->avg.util_avg; in update_tg_cfs_util()
3729 new_sum = se->avg.util_avg * divider; in update_tg_cfs_util()
3730 delta_sum = (long)new_sum - (long)se->avg.util_sum; in update_tg_cfs_util()
3731 se->avg.util_sum = new_sum; in update_tg_cfs_util()
3734 add_positive(&cfs_rq->avg.util_avg, delta_avg); in update_tg_cfs_util()
3735 add_positive(&cfs_rq->avg.util_sum, delta_sum); in update_tg_cfs_util()
3738 cfs_rq->avg.util_sum = max_t(u32, cfs_rq->avg.util_sum, in update_tg_cfs_util()
3739 cfs_rq->avg.util_avg * PELT_MIN_DIVIDER); in update_tg_cfs_util()
3745 long delta_sum, delta_avg = gcfs_rq->avg.runnable_avg - se->avg.runnable_avg; in update_tg_cfs_runnable()
3756 divider = get_pelt_divider(&cfs_rq->avg); in update_tg_cfs_runnable()
3759 se->avg.runnable_avg = gcfs_rq->avg.runnable_avg; in update_tg_cfs_runnable()
3760 new_sum = se->avg.runnable_avg * divider; in update_tg_cfs_runnable()
3761 delta_sum = (long)new_sum - (long)se->avg.runnable_sum; in update_tg_cfs_runnable()
3762 se->avg.runnable_sum = new_sum; in update_tg_cfs_runnable()
3765 add_positive(&cfs_rq->avg.runnable_avg, delta_avg); in update_tg_cfs_runnable()
3766 add_positive(&cfs_rq->avg.runnable_sum, delta_sum); in update_tg_cfs_runnable()
3768 cfs_rq->avg.runnable_sum = max_t(u32, cfs_rq->avg.runnable_sum, in update_tg_cfs_runnable()
3769 cfs_rq->avg.runnable_avg * PELT_MIN_DIVIDER); in update_tg_cfs_runnable()
3790 divider = get_pelt_divider(&cfs_rq->avg); in update_tg_cfs_load()
3797 runnable_sum += se->avg.load_sum; in update_tg_cfs_load()
3805 load_sum = div_u64(gcfs_rq->avg.load_sum, in update_tg_cfs_load()
3810 runnable_sum = min(se->avg.load_sum, load_sum); in update_tg_cfs_load()
3819 running_sum = se->avg.util_sum >> SCHED_CAPACITY_SHIFT; in update_tg_cfs_load()
3825 delta_avg = load_avg - se->avg.load_avg; in update_tg_cfs_load()
3829 delta_sum = load_sum - (s64)se_weight(se) * se->avg.load_sum; in update_tg_cfs_load()
3831 se->avg.load_sum = runnable_sum; in update_tg_cfs_load()
3832 se->avg.load_avg = load_avg; in update_tg_cfs_load()
3833 add_positive(&cfs_rq->avg.load_avg, delta_avg); in update_tg_cfs_load()
3834 add_positive(&cfs_rq->avg.load_sum, delta_sum); in update_tg_cfs_load()
3836 cfs_rq->avg.load_sum = max_t(u32, cfs_rq->avg.load_sum, in update_tg_cfs_load()
3837 cfs_rq->avg.load_avg * PELT_MIN_DIVIDER); in update_tg_cfs_load()
3886 if (se->avg.load_avg || se->avg.util_avg) in skip_blocked_update()
3925 if (load_avg_is_decayed(&se->avg)) in migrate_se_pelt_lag()
4019 struct sched_avg *sa = &cfs_rq->avg; in update_cfs_rq_load_avg()
4024 u32 divider = get_pelt_divider(&cfs_rq->avg); in update_cfs_rq_load_avg()
4093 u32 divider = get_pelt_divider(&cfs_rq->avg); in attach_entity_load_avg()
4102 se->avg.last_update_time = cfs_rq->avg.last_update_time; in attach_entity_load_avg()
4103 se->avg.period_contrib = cfs_rq->avg.period_contrib; in attach_entity_load_avg()
4111 se->avg.util_sum = se->avg.util_avg * divider; in attach_entity_load_avg()
4113 se->avg.runnable_sum = se->avg.runnable_avg * divider; in attach_entity_load_avg()
4115 se->avg.load_sum = se->avg.load_avg * divider; in attach_entity_load_avg()
4116 if (se_weight(se) < se->avg.load_sum) in attach_entity_load_avg()
4117 se->avg.load_sum = div_u64(se->avg.load_sum, se_weight(se)); in attach_entity_load_avg()
4119 se->avg.load_sum = 1; in attach_entity_load_avg()
4122 cfs_rq->avg.util_avg += se->avg.util_avg; in attach_entity_load_avg()
4123 cfs_rq->avg.util_sum += se->avg.util_sum; in attach_entity_load_avg()
4124 cfs_rq->avg.runnable_avg += se->avg.runnable_avg; in attach_entity_load_avg()
4125 cfs_rq->avg.runnable_sum += se->avg.runnable_sum; in attach_entity_load_avg()
4127 add_tg_cfs_propagate(cfs_rq, se->avg.load_sum); in attach_entity_load_avg()
4145 sub_positive(&cfs_rq->avg.util_avg, se->avg.util_avg); in detach_entity_load_avg()
4146 sub_positive(&cfs_rq->avg.util_sum, se->avg.util_sum); in detach_entity_load_avg()
4148 cfs_rq->avg.util_sum = max_t(u32, cfs_rq->avg.util_sum, in detach_entity_load_avg()
4149 cfs_rq->avg.util_avg * PELT_MIN_DIVIDER); in detach_entity_load_avg()
4151 sub_positive(&cfs_rq->avg.runnable_avg, se->avg.runnable_avg); in detach_entity_load_avg()
4152 sub_positive(&cfs_rq->avg.runnable_sum, se->avg.runnable_sum); in detach_entity_load_avg()
4154 cfs_rq->avg.runnable_sum = max_t(u32, cfs_rq->avg.runnable_sum, in detach_entity_load_avg()
4155 cfs_rq->avg.runnable_avg * PELT_MIN_DIVIDER); in detach_entity_load_avg()
4157 add_tg_cfs_propagate(cfs_rq, -se->avg.load_sum); in detach_entity_load_avg()
4182 if (se->avg.last_update_time && !(flags & SKIP_AGE_LOAD)) in update_load_avg()
4188 if (!se->avg.last_update_time && (flags & DO_ATTACH)) { in update_load_avg()
4247 cfs_rq->removed.util_avg += se->avg.util_avg; in remove_entity_load_avg()
4248 cfs_rq->removed.load_avg += se->avg.load_avg; in remove_entity_load_avg()
4249 cfs_rq->removed.runnable_avg += se->avg.runnable_avg; in remove_entity_load_avg()
4255 return cfs_rq->avg.runnable_avg; in cfs_rq_runnable_avg()
4260 return cfs_rq->avg.load_avg; in cfs_rq_load_avg()
4267 return READ_ONCE(p->se.avg.util_avg); in task_util()
4272 struct util_est ue = READ_ONCE(p->se.avg.util_est); in _task_util_est()
4305 enqueued = cfs_rq->avg.util_est.enqueued; in util_est_enqueue()
4307 WRITE_ONCE(cfs_rq->avg.util_est.enqueued, enqueued); in util_est_enqueue()
4321 enqueued = cfs_rq->avg.util_est.enqueued; in util_est_dequeue()
4323 WRITE_ONCE(cfs_rq->avg.util_est.enqueued, enqueued); in util_est_dequeue()
4364 ue = p->se.avg.util_est; in util_est_update()
4424 WRITE_ONCE(p->se.avg.util_est, ue); in util_est_update()
6116 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_load_without()
6120 load = READ_ONCE(cfs_rq->avg.load_avg); in cpu_load_without()
6139 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_runnable_without()
6143 runnable = READ_ONCE(cfs_rq->avg.runnable_avg); in cpu_runnable_without()
6146 lsub_positive(&runnable, p->se.avg.runnable_avg); in cpu_runnable_without()
6803 unsigned long util = READ_ONCE(cfs_rq->avg.util_avg); in cpu_util_next()
6819 util_est = READ_ONCE(cfs_rq->avg.util_est.enqueued); in cpu_util_next()
6874 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_util_without()
7293 se->avg.last_update_time = 0; in migrate_task_rq_fair()
8375 if (cfs_rq->avg.load_avg) in cfs_rq_has_blocked()
8378 if (cfs_rq->avg.util_avg) in cfs_rq_has_blocked()
8521 load = div64_ul(load * se->avg.load_avg, in update_cfs_rq_h_load()
8534 return div64_ul(p->se.avg.load_avg * cfs_rq->h_load, in task_h_load()
8552 return p->se.avg.load_avg; in task_h_load()
9231 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in task_running_on_cpu()
11730 if (!se->avg.last_update_time) in detach_entity_cfs_rq()
11853 p->se.avg.last_update_time = 0; in task_change_group_fair()