Lines Matching refs:sa
110 accumulate_sum(u64 delta, struct sched_avg *sa, in accumulate_sum() argument
116 delta += sa->period_contrib; in accumulate_sum()
123 sa->load_sum = decay_load(sa->load_sum, periods); in accumulate_sum()
124 sa->runnable_load_sum = in accumulate_sum()
125 decay_load(sa->runnable_load_sum, periods); in accumulate_sum()
126 sa->util_sum = decay_load((u64)(sa->util_sum), periods); in accumulate_sum()
133 1024 - sa->period_contrib, delta); in accumulate_sum()
135 sa->period_contrib = delta; in accumulate_sum()
138 sa->load_sum += load * contrib; in accumulate_sum()
140 sa->runnable_load_sum += runnable * contrib; in accumulate_sum()
142 sa->util_sum += contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum()
176 ___update_load_sum(u64 now, struct sched_avg *sa, in ___update_load_sum() argument
181 delta = now - sa->last_update_time; in ___update_load_sum()
187 sa->last_update_time = now; in ___update_load_sum()
199 sa->last_update_time += delta << 10; in ___update_load_sum()
220 if (!accumulate_sum(delta, sa, load, runnable, running)) in ___update_load_sum()
227 ___update_load_avg(struct sched_avg *sa, unsigned long load, unsigned long runnable) in ___update_load_avg() argument
229 u32 divider = LOAD_AVG_MAX - 1024 + sa->period_contrib; in ___update_load_avg()
234 sa->load_avg = div_u64(load * sa->load_sum, divider); in ___update_load_avg()
235 sa->runnable_load_avg = div_u64(runnable * sa->runnable_load_sum, divider); in ___update_load_avg()
236 WRITE_ONCE(sa->util_avg, sa->util_sum / divider); in ___update_load_avg()