Lines Matching refs:avg
700 struct sched_avg *sa = &se->avg; in init_entity_runnable_average()
750 struct sched_avg *sa = &se->avg; in post_init_entity_util_avg()
752 long cap = (long)(cpu_scale - cfs_rq->avg.util_avg) / 2; in post_init_entity_util_avg()
755 if (cfs_rq->avg.util_avg != 0) { in post_init_entity_util_avg()
756 sa->util_avg = cfs_rq->avg.util_avg * se->load.weight; in post_init_entity_util_avg()
757 sa->util_avg /= (cfs_rq->avg.load_avg + 1); in post_init_entity_util_avg()
779 se->avg.last_update_time = cfs_rq_clock_task(cfs_rq); in post_init_entity_util_avg()
2021 delta = p->se.avg.load_sum; in numa_get_avg_runtime()
2765 cfs_rq->avg.runnable_load_avg += se->avg.runnable_load_avg; in enqueue_runnable_load_avg()
2766 cfs_rq->avg.runnable_load_sum += se_runnable(se) * se->avg.runnable_load_sum; in enqueue_runnable_load_avg()
2774 sub_positive(&cfs_rq->avg.runnable_load_avg, se->avg.runnable_load_avg); in dequeue_runnable_load_avg()
2775 sub_positive(&cfs_rq->avg.runnable_load_sum, in dequeue_runnable_load_avg()
2776 se_runnable(se) * se->avg.runnable_load_sum); in dequeue_runnable_load_avg()
2782 cfs_rq->avg.load_avg += se->avg.load_avg; in enqueue_load_avg()
2783 cfs_rq->avg.load_sum += se_weight(se) * se->avg.load_sum; in enqueue_load_avg()
2789 sub_positive(&cfs_rq->avg.load_avg, se->avg.load_avg); in dequeue_load_avg()
2790 sub_positive(&cfs_rq->avg.load_sum, se_weight(se) * se->avg.load_sum); in dequeue_load_avg()
2820 u32 divider = LOAD_AVG_MAX - 1024 + se->avg.period_contrib; in reweight_entity()
2822 se->avg.load_avg = div_u64(se_weight(se) * se->avg.load_sum, divider); in reweight_entity()
2823 se->avg.runnable_load_avg = in reweight_entity()
2824 div_u64(se_runnable(se) * se->avg.runnable_load_sum, divider); in reweight_entity()
2928 load = max(scale_load_down(cfs_rq->load.weight), cfs_rq->avg.load_avg); in calc_group_shares()
2986 load_avg = max(cfs_rq->avg.load_avg, in calc_group_runnable()
2989 runnable = max(cfs_rq->avg.runnable_load_avg, in calc_group_runnable()
3078 long delta = cfs_rq->avg.load_avg - cfs_rq->tg_load_avg_contrib; in update_tg_load_avg()
3088 cfs_rq->tg_load_avg_contrib = cfs_rq->avg.load_avg; in update_tg_load_avg()
3113 if (!(se->avg.last_update_time && prev)) in set_task_rq_fair()
3127 p_last_update_time = prev->avg.last_update_time; in set_task_rq_fair()
3128 n_last_update_time = next->avg.last_update_time; in set_task_rq_fair()
3134 p_last_update_time = prev->avg.last_update_time; in set_task_rq_fair()
3135 n_last_update_time = next->avg.last_update_time; in set_task_rq_fair()
3138 se->avg.last_update_time = n_last_update_time; in set_task_rq_fair()
3213 long delta = gcfs_rq->avg.util_avg - se->avg.util_avg; in update_tg_cfs_util()
3228 se->avg.util_avg = gcfs_rq->avg.util_avg; in update_tg_cfs_util()
3229 se->avg.util_sum = se->avg.util_avg * LOAD_AVG_MAX; in update_tg_cfs_util()
3232 add_positive(&cfs_rq->avg.util_avg, delta); in update_tg_cfs_util()
3233 cfs_rq->avg.util_sum = cfs_rq->avg.util_avg * LOAD_AVG_MAX; in update_tg_cfs_util()
3254 runnable_sum += se->avg.load_sum; in update_tg_cfs_runnable()
3262 load_sum = div_s64(gcfs_rq->avg.load_sum, in update_tg_cfs_runnable()
3267 runnable_sum = min(se->avg.load_sum, load_sum); in update_tg_cfs_runnable()
3275 running_sum = se->avg.util_sum / in update_tg_cfs_runnable()
3282 delta_sum = load_sum - (s64)se_weight(se) * se->avg.load_sum; in update_tg_cfs_runnable()
3283 delta_avg = load_avg - se->avg.load_avg; in update_tg_cfs_runnable()
3285 se->avg.load_sum = runnable_sum; in update_tg_cfs_runnable()
3286 se->avg.load_avg = load_avg; in update_tg_cfs_runnable()
3287 add_positive(&cfs_rq->avg.load_avg, delta_avg); in update_tg_cfs_runnable()
3288 add_positive(&cfs_rq->avg.load_sum, delta_sum); in update_tg_cfs_runnable()
3292 delta_sum = runnable_load_sum - se_weight(se) * se->avg.runnable_load_sum; in update_tg_cfs_runnable()
3293 delta_avg = runnable_load_avg - se->avg.runnable_load_avg; in update_tg_cfs_runnable()
3295 se->avg.runnable_load_sum = runnable_sum; in update_tg_cfs_runnable()
3296 se->avg.runnable_load_avg = runnable_load_avg; in update_tg_cfs_runnable()
3299 add_positive(&cfs_rq->avg.runnable_load_avg, delta_avg); in update_tg_cfs_runnable()
3300 add_positive(&cfs_rq->avg.runnable_load_sum, delta_sum); in update_tg_cfs_runnable()
3346 if (se->avg.load_avg || se->avg.util_avg) in skip_blocked_update()
3397 struct sched_avg *sa = &cfs_rq->avg; in update_cfs_rq_load_avg()
3448 u32 divider = LOAD_AVG_MAX - 1024 + cfs_rq->avg.period_contrib; in attach_entity_load_avg()
3457 se->avg.last_update_time = cfs_rq->avg.last_update_time; in attach_entity_load_avg()
3458 se->avg.period_contrib = cfs_rq->avg.period_contrib; in attach_entity_load_avg()
3466 se->avg.util_sum = se->avg.util_avg * divider; in attach_entity_load_avg()
3468 se->avg.load_sum = divider; in attach_entity_load_avg()
3470 se->avg.load_sum = in attach_entity_load_avg()
3471 div_u64(se->avg.load_avg * se->avg.load_sum, se_weight(se)); in attach_entity_load_avg()
3474 se->avg.runnable_load_sum = se->avg.load_sum; in attach_entity_load_avg()
3477 cfs_rq->avg.util_avg += se->avg.util_avg; in attach_entity_load_avg()
3478 cfs_rq->avg.util_sum += se->avg.util_sum; in attach_entity_load_avg()
3480 add_tg_cfs_propagate(cfs_rq, se->avg.load_sum); in attach_entity_load_avg()
3496 sub_positive(&cfs_rq->avg.util_avg, se->avg.util_avg); in detach_entity_load_avg()
3497 sub_positive(&cfs_rq->avg.util_sum, se->avg.util_sum); in detach_entity_load_avg()
3499 add_tg_cfs_propagate(cfs_rq, -se->avg.load_sum); in detach_entity_load_avg()
3523 if (se->avg.last_update_time && !(flags & SKIP_AGE_LOAD)) in update_load_avg()
3529 if (!se->avg.last_update_time && (flags & DO_ATTACH)) { in update_load_avg()
3554 last_update_time = cfs_rq->avg.last_update_time; in cfs_rq_last_update_time()
3562 return cfs_rq->avg.last_update_time; in cfs_rq_last_update_time()
3602 cfs_rq->removed.util_avg += se->avg.util_avg; in remove_entity_load_avg()
3603 cfs_rq->removed.load_avg += se->avg.load_avg; in remove_entity_load_avg()
3604 cfs_rq->removed.runnable_sum += se->avg.load_sum; /* == runnable_sum */ in remove_entity_load_avg()
3610 return cfs_rq->avg.runnable_load_avg; in cfs_rq_runnable_load_avg()
3615 return cfs_rq->avg.load_avg; in cfs_rq_load_avg()
3622 return READ_ONCE(p->se.avg.util_avg); in task_util()
3627 struct util_est ue = READ_ONCE(p->se.avg.util_est); in _task_util_est()
3646 enqueued = cfs_rq->avg.util_est.enqueued; in util_est_enqueue()
3648 WRITE_ONCE(cfs_rq->avg.util_est.enqueued, enqueued); in util_est_enqueue()
3674 ue.enqueued = cfs_rq->avg.util_est.enqueued; in util_est_dequeue()
3677 WRITE_ONCE(cfs_rq->avg.util_est.enqueued, ue.enqueued); in util_est_dequeue()
3690 ue = p->se.avg.util_est; in util_est_dequeue()
3723 WRITE_ONCE(p->se.avg.util_est, ue); in util_est_dequeue()
6208 util = READ_ONCE(cfs_rq->avg.util_avg); in cpu_util()
6211 util = max(util, READ_ONCE(cfs_rq->avg.util_est.enqueued)); in cpu_util()
6226 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_util_wake()
6230 util = READ_ONCE(cfs_rq->avg.util_avg); in cpu_util_wake()
6262 util = max(util, READ_ONCE(cfs_rq->avg.util_est.enqueued)); in cpu_util_wake()
6419 p->se.avg.last_update_time = 0; in migrate_task_rq_fair()
7313 if (cfs_rq->avg.load_avg) in cfs_rq_has_blocked()
7316 if (cfs_rq->avg.util_avg) in cfs_rq_has_blocked()
7345 if (cfs_rq->avg.load_sum) in cfs_rq_is_decayed()
7348 if (cfs_rq->avg.util_sum) in cfs_rq_is_decayed()
7351 if (cfs_rq->avg.runnable_load_sum) in cfs_rq_is_decayed()
7445 load = div64_ul(load * se->avg.load_avg, in update_cfs_rq_h_load()
7458 return div64_ul(p->se.avg.load_avg * cfs_rq->h_load, in task_h_load()
7487 return p->se.avg.load_avg; in task_h_load()
9900 p->se.avg.last_update_time = 0; in task_move_group_fair()