Lines Matching refs:prev_cpu

694 static int select_idle_sibling(struct task_struct *p, int prev_cpu, int cpu);
5589 wake_affine_idle(int this_cpu, int prev_cpu, int sync) in wake_affine_idle() argument
5603 if (available_idle_cpu(this_cpu) && cpus_share_cache(this_cpu, prev_cpu)) in wake_affine_idle()
5604 return available_idle_cpu(prev_cpu) ? prev_cpu : this_cpu; in wake_affine_idle()
5614 int this_cpu, int prev_cpu, int sync) in wake_affine_weight() argument
5635 this_eff_load *= capacity_of(prev_cpu); in wake_affine_weight()
5637 prev_eff_load = source_load(prev_cpu, sd->wake_idx); in wake_affine_weight()
5656 int this_cpu, int prev_cpu, int sync) in wake_affine() argument
5661 target = wake_affine_idle(this_cpu, prev_cpu, sync); in wake_affine()
5664 target = wake_affine_weight(sd, p, this_cpu, prev_cpu, sync); in wake_affine()
5668 return prev_cpu; in wake_affine()
5882 int cpu, int prev_cpu, int sd_flag) in find_idlest_cpu() argument
5887 return prev_cpu; in find_idlest_cpu()
6279 static int wake_cap(struct task_struct *p, int cpu, int prev_cpu) in wake_cap() argument
6283 min_cap = min(capacity_orig_of(prev_cpu), capacity_orig_of(cpu)); in wake_cap()
6309 select_task_rq_fair(struct task_struct *p, int prev_cpu, int sd_flag, int wake_flags) in select_task_rq_fair() argument
6313 int new_cpu = prev_cpu; in select_task_rq_fair()
6319 want_affine = !wake_wide(p) && !wake_cap(p, cpu, prev_cpu) in select_task_rq_fair()
6333 cpumask_test_cpu(prev_cpu, sched_domain_span(tmp))) { in select_task_rq_fair()
6334 if (cpu != prev_cpu) in select_task_rq_fair()
6335 new_cpu = wake_affine(tmp, p, cpu, prev_cpu, sync); in select_task_rq_fair()
6349 new_cpu = find_idlest_cpu(sd, p, cpu, prev_cpu, sd_flag); in select_task_rq_fair()
6353 new_cpu = select_idle_sibling(p, prev_cpu, new_cpu); in select_task_rq_fair()