Lines Matching refs:rq

7 int update_rt_rq_load_avg(u64 now, struct rq *rq, int running);
8 int update_dl_rq_load_avg(u64 now, struct rq *rq, int running);
11 int update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity);
13 static inline u64 thermal_load_avg(struct rq *rq) in thermal_load_avg() argument
15 return READ_ONCE(rq->avg_thermal.load_avg); in thermal_load_avg()
19 update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity) in update_thermal_load_avg() argument
24 static inline u64 thermal_load_avg(struct rq *rq) in thermal_load_avg() argument
31 int update_irq_load_avg(struct rq *rq, u64 running);
34 update_irq_load_avg(struct rq *rq, u64 running) in update_irq_load_avg() argument
64 static inline u64 rq_clock_pelt(struct rq *rq) in rq_clock_pelt() argument
66 lockdep_assert_rq_held(rq); in rq_clock_pelt()
67 assert_clock_updated(rq); in rq_clock_pelt()
69 return rq->clock_pelt - rq->lost_idle_time; in rq_clock_pelt()
73 static inline void _update_idle_rq_clock_pelt(struct rq *rq) in _update_idle_rq_clock_pelt() argument
75 rq->clock_pelt = rq_clock_task(rq); in _update_idle_rq_clock_pelt()
77 u64_u32_store(rq->clock_idle, rq_clock(rq)); in _update_idle_rq_clock_pelt()
80 u64_u32_store(rq->clock_pelt_idle, rq_clock_pelt(rq)); in _update_idle_rq_clock_pelt()
95 static inline void update_rq_clock_pelt(struct rq *rq, s64 delta) in update_rq_clock_pelt() argument
97 if (unlikely(is_idle_task(rq->curr))) { in update_rq_clock_pelt()
98 _update_idle_rq_clock_pelt(rq); in update_rq_clock_pelt()
118 delta = cap_scale(delta, arch_scale_cpu_capacity(cpu_of(rq))); in update_rq_clock_pelt()
119 delta = cap_scale(delta, arch_scale_freq_capacity(cpu_of(rq))); in update_rq_clock_pelt()
121 rq->clock_pelt += delta; in update_rq_clock_pelt()
133 static inline void update_idle_rq_clock_pelt(struct rq *rq) in update_idle_rq_clock_pelt() argument
136 u32 util_sum = rq->cfs.avg.util_sum; in update_idle_rq_clock_pelt()
137 util_sum += rq->avg_rt.util_sum; in update_idle_rq_clock_pelt()
138 util_sum += rq->avg_dl.util_sum; in update_idle_rq_clock_pelt()
150 rq->lost_idle_time += rq_clock_task(rq) - rq->clock_pelt; in update_idle_rq_clock_pelt()
152 _update_idle_rq_clock_pelt(rq); in update_idle_rq_clock_pelt()
193 update_rt_rq_load_avg(u64 now, struct rq *rq, int running) in update_rt_rq_load_avg() argument
199 update_dl_rq_load_avg(u64 now, struct rq *rq, int running) in update_dl_rq_load_avg() argument
205 update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity) in update_thermal_load_avg() argument
210 static inline u64 thermal_load_avg(struct rq *rq) in thermal_load_avg() argument
216 update_irq_load_avg(struct rq *rq, u64 running) in update_irq_load_avg() argument
221 static inline u64 rq_clock_pelt(struct rq *rq) in rq_clock_pelt() argument
223 return rq_clock_task(rq); in rq_clock_pelt()
227 update_rq_clock_pelt(struct rq *rq, s64 delta) { } in update_rq_clock_pelt() argument
230 update_idle_rq_clock_pelt(struct rq *rq) { } in update_idle_rq_clock_pelt() argument