Home
last modified time | relevance | path

Searched refs:cpu_of (Results 1 – 7 of 7) sorted by relevance

/Linux-v5.15/kernel/sched/
Dpelt.h98 delta = cap_scale(delta, arch_scale_cpu_capacity(cpu_of(rq))); in update_rq_clock_pelt()
99 delta = cap_scale(delta, arch_scale_freq_capacity(cpu_of(rq))); in update_rq_clock_pelt()
Dpelt.c443 running = cap_scale(running, arch_scale_freq_capacity(cpu_of(rq))); in update_irq_load_avg()
444 running = cap_scale(running, arch_scale_cpu_capacity(cpu_of(rq))); in update_irq_load_avg()
Dfair.c292 int cpu = cpu_of(rq); in list_add_leaf_cfs_rq()
775 long cpu_scale = arch_scale_cpu_capacity(cpu_of(rq_of(cfs_rq))); in post_init_entity_util_avg()
4083 if (task_util(p) > capacity_orig_of(cpu_of(rq_of(cfs_rq)))) in util_est_update()
4128 if (task_fits_capacity(p, capacity_of(cpu_of(rq)))) { in update_misfit_status()
4801 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_unthrottle_up()
4819 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in tg_throttle_down()
4859 se = cfs_rq->tg->se[cpu_of(rq_of(cfs_rq))]; in throttle_cfs_rq()
4925 se = cfs_rq->tg->se[cpu_of(rq)]; in unthrottle_cfs_rq()
5417 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in update_runtime_enabled()
5435 struct cfs_rq *cfs_rq = tg->cfs_rq[cpu_of(rq)]; in unthrottle_offline_cfs_rqs()
[all …]
Dsched.h1131 static inline int cpu_of(struct rq *rq) in cpu_of() function
1213 for_each_cpu(cpu, cpu_smt_mask(cpu_of(rq))) { in sched_core_cookie_match()
2349 int cpu = cpu_of(rq); in sched_update_tick_dependency()
2426 if (!cpu_active(cpu_of(rq))) in hrtick_enabled()
2818 cpu_of(rq))); in cpufreq_update_util()
Dcore.c619 irq_delta = irq_time_read(cpu_of(rq)) - rq->prev_irq_time; in update_rq_clock_task()
644 steal = paravirt_steal_clock(cpu_of(rq)); in update_rq_clock_task()
679 delta = sched_clock_cpu(cpu_of(rq)) - rq->clock; in update_rq_clock()
706 WARN_ON_ONCE(cpu_of(rq) != smp_processor_id()); in hrtick()
759 smp_call_function_single_async(cpu_of(rq), &rq->hrtick_csd); in hrtick_start()
971 cpu = cpu_of(rq); in resched_curr()
1098 int cpu = cpu_of(rq); in nohz_csd_func()
2704 stop_one_cpu_nowait(cpu_of(rq), migration_cpu_stop, in affine_move_task()
3649 if (WARN_ON_ONCE(task_cpu(p) != cpu_of(rq))) in sched_ttwu_pending()
3650 set_task_cpu(p, cpu_of(rq)); in sched_ttwu_pending()
[all …]
Drt.c515 (rt_rq = iter->rt_rq[cpu_of(rq)]);)
534 int cpu = cpu_of(rq); in sched_rt_rq_enqueue()
552 int cpu = cpu_of(rq_of_rt_rq(rt_rq)); in sched_rt_rq_dequeue()
2362 if (p->prio < rq->curr->prio && cpu_online(cpu_of(rq))) in switched_to_rt()
Ddeadline.c1246 int cpu = cpu_of(rq); in update_curr_dl()
2353 src_dl_b = dl_bw_of(cpu_of(rq)); in set_cpus_allowed_dl()