Lines Matching refs:dst_cpu
1388 int src_nid, int dst_cpu) in should_numa_migrate_memory() argument
1391 int dst_nid = cpu_to_node(dst_cpu); in should_numa_migrate_memory()
1394 this_cpupid = cpu_pid_to_cpupid(dst_cpu, current->pid); in should_numa_migrate_memory()
1512 int dst_cpu, dst_nid; member
1527 struct rq *rq = cpu_rq(env->dst_cpu); in task_numa_assign()
1549 env->best_cpu = env->dst_cpu; in task_numa_assign()
1596 struct rq *dst_rq = cpu_rq(env->dst_cpu); in task_numa_compare()
1702 env->dst_cpu = select_idle_sibling(env->p, env->src_cpu, in task_numa_compare()
1703 env->dst_cpu); in task_numa_compare()
1734 env->dst_cpu = cpu; in task_numa_find_cpu()
6933 int dst_cpu; member
7005 dst_nid = cpu_to_node(env->dst_cpu); in migrate_degrades_locality()
7063 if (throttled_lb_pair(task_group(p), env->src_cpu, env->dst_cpu)) in can_migrate_task()
7066 if (!cpumask_test_cpu(env->dst_cpu, &p->cpus_allowed)) { in can_migrate_task()
7136 set_task_cpu(p, env->dst_cpu); in detach_task()
7951 sched_asym_prefer(env->dst_cpu, sg->asym_prefer_cpu)) { in update_sd_pick_busiest()
8022 local_group = cpumask_test_cpu(env->dst_cpu, sched_group_span(sg)); in update_sd_lb_stats()
8029 update_group_capacity(env->sd, env->dst_cpu); in update_sd_lb_stats()
8125 if (sched_asym_prefer(busiest_cpu, env->dst_cpu)) in check_asym_packing()
8154 local->load_per_task = cpu_avg_load_per_task(env->dst_cpu); in fix_small_imbalance()
8471 sched_asym_prefer(env->dst_cpu, env->src_cpu)) in need_active_balance()
8484 (capacity_of(env->src_cpu)*sd->imbalance_pct < capacity_of(env->dst_cpu)*100)) in need_active_balance()
8502 if (!cpumask_test_cpu(env->dst_cpu, env->cpus)) in should_we_balance()
8528 return balance_cpu == env->dst_cpu; in should_we_balance()
8548 .dst_cpu = this_cpu, in load_balance()
8652 cpumask_clear_cpu(env.dst_cpu, env.cpus); in load_balance()
8655 env.dst_cpu = env.new_dst_cpu; in load_balance()
8879 .dst_cpu = target_cpu, in active_load_balance_cpu_stop()