/Linux-v4.19/arch/ia64/kernel/ |
D | irq.c | 104 int irq, new_cpu; in migrate_irqs() local 130 new_cpu = cpumask_any(cpu_online_mask); in migrate_irqs() 139 cpumask_of(new_cpu), false); in migrate_irqs()
|
/Linux-v4.19/drivers/irqchip/ |
D | irq-bcm6345-l1.c | 207 unsigned int new_cpu; in bcm6345_l1_set_affinity() local 215 new_cpu = cpumask_any_and(&valid, cpu_online_mask); in bcm6345_l1_set_affinity() 216 if (new_cpu >= nr_cpu_ids) in bcm6345_l1_set_affinity() 219 dest = cpumask_of(new_cpu); in bcm6345_l1_set_affinity() 222 if (old_cpu != new_cpu) { in bcm6345_l1_set_affinity() 234 irq_data_update_effective_affinity(d, cpumask_of(new_cpu)); in bcm6345_l1_set_affinity()
|
/Linux-v4.19/kernel/sched/ |
D | cpudeadline.c | 156 int old_idx, new_cpu; in cpudl_clear() local 171 new_cpu = cp->elements[cp->size - 1].cpu; in cpudl_clear() 173 cp->elements[old_idx].cpu = new_cpu; in cpudl_clear() 175 cp->elements[new_cpu].idx = old_idx; in cpudl_clear()
|
D | fair.c | 2638 static void update_scan_period(struct task_struct *p, int new_cpu) in update_scan_period() argument 2641 int dst_nid = cpu_to_node(new_cpu); in update_scan_period() 2684 static inline void update_scan_period(struct task_struct *p, int new_cpu) in update_scan_period() argument 5884 int new_cpu = cpu; in find_idlest_cpu() local 5912 new_cpu = find_idlest_group_cpu(group, p, cpu); in find_idlest_cpu() 5913 if (new_cpu == cpu) { in find_idlest_cpu() 5920 cpu = new_cpu; in find_idlest_cpu() 5931 return new_cpu; in find_idlest_cpu() 6313 int new_cpu = prev_cpu; in select_task_rq_fair() local 6335 new_cpu = wake_affine(tmp, p, cpu, prev_cpu, sync); in select_task_rq_fair() [all …]
|
D | core.c | 909 struct task_struct *p, int new_cpu) in move_queued_task() argument 915 set_task_cpu(p, new_cpu); in move_queued_task() 918 rq = cpu_rq(new_cpu); in move_queued_task() 921 BUG_ON(task_cpu(p) != new_cpu); in move_queued_task() 1127 void set_task_cpu(struct task_struct *p, unsigned int new_cpu) in set_task_cpu() argument 1163 WARN_ON_ONCE(!cpu_online(new_cpu)); in set_task_cpu() 1166 trace_sched_migrate_task(p, new_cpu); in set_task_cpu() 1168 if (task_cpu(p) != new_cpu) { in set_task_cpu() 1170 p->sched_class->migrate_task_rq(p, new_cpu); in set_task_cpu() 1176 __set_task_cpu(p, new_cpu); in set_task_cpu()
|
D | sched.h | 1529 void (*migrate_task_rq)(struct task_struct *p, int new_cpu);
|
D | deadline.c | 1610 static void migrate_task_rq_dl(struct task_struct *p, int new_cpu __maybe_unused) in migrate_task_rq_dl()
|
/Linux-v4.19/arch/x86/hyperv/ |
D | hv_init.c | 226 unsigned int new_cpu; in hv_cpu_die() local 247 new_cpu = cpumask_any_but(cpu_online_mask, cpu); in hv_cpu_die() 249 re_ctrl.target_vp = hv_vp_index[new_cpu]; in hv_cpu_die()
|
/Linux-v4.19/tools/perf/scripts/python/ |
D | sched-migration.py | 193 def migrate(self, ts_list, new, old_cpu, new_cpu): argument 194 if old_cpu == new_cpu: 201 new_rq = self.prev.rqs[new_cpu] 203 self.rqs[new_cpu] = in_rq 210 self.event_cpus.append(new_cpu)
|
/Linux-v4.19/arch/x86/events/intel/ |
D | uncore.c | 1113 int new_cpu) in uncore_change_type_ctx() argument 1119 pkg = topology_logical_package_id(old_cpu < 0 ? new_cpu : old_cpu); in uncore_change_type_ctx() 1127 box->cpu = new_cpu; in uncore_change_type_ctx() 1133 if (new_cpu < 0) in uncore_change_type_ctx() 1137 perf_pmu_migrate_context(&pmu->pmu, old_cpu, new_cpu); in uncore_change_type_ctx() 1138 box->cpu = new_cpu; in uncore_change_type_ctx() 1143 int old_cpu, int new_cpu) in uncore_change_context() argument 1146 uncore_change_type_ctx(*uncores, old_cpu, new_cpu); in uncore_change_context()
|
/Linux-v4.19/virt/kvm/arm/vgic/ |
D | vgic.c | 670 struct vgic_cpu *new_cpu = &target_vcpu->arch.vgic_cpu; in vgic_prune_ap_list() local 674 list_add_tail(&irq->ap_list, &new_cpu->ap_list_head); in vgic_prune_ap_list()
|
/Linux-v4.19/arch/powerpc/perf/ |
D | imc-pmu.c | 300 static void nest_change_cpu_context(int old_cpu, int new_cpu) in nest_change_cpu_context() argument 304 if (old_cpu < 0 || new_cpu < 0) in nest_change_cpu_context() 308 perf_pmu_migrate_context(&(*pn)->pmu, old_cpu, new_cpu); in nest_change_cpu_context()
|
/Linux-v4.19/kernel/ |
D | workqueue.c | 1338 int new_cpu; in wq_select_unbound_cpu() local 1351 new_cpu = __this_cpu_read(wq_rr_cpu_last); in wq_select_unbound_cpu() 1352 new_cpu = cpumask_next_and(new_cpu, wq_unbound_cpumask, cpu_online_mask); in wq_select_unbound_cpu() 1353 if (unlikely(new_cpu >= nr_cpu_ids)) { in wq_select_unbound_cpu() 1354 new_cpu = cpumask_first_and(wq_unbound_cpumask, cpu_online_mask); in wq_select_unbound_cpu() 1355 if (unlikely(new_cpu >= nr_cpu_ids)) in wq_select_unbound_cpu() 1358 __this_cpu_write(wq_rr_cpu_last, new_cpu); in wq_select_unbound_cpu() 1360 return new_cpu; in wq_select_unbound_cpu()
|
/Linux-v4.19/tools/perf/ |
D | builtin-sched.c | 1524 bool new_cpu = false; in map_switch_event() local 1537 new_cpu = true; in map_switch_event() 1642 if (sched->map.comp && new_cpu) in map_switch_event()
|