/Linux-v5.10/include/linux/ |
D | cpumask.h | 17 typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t; struct 90 extern struct cpumask __cpu_possible_mask; 91 extern struct cpumask __cpu_online_mask; 92 extern struct cpumask __cpu_present_mask; 93 extern struct cpumask __cpu_active_mask; 94 #define cpu_possible_mask ((const struct cpumask *)&__cpu_possible_mask) 95 #define cpu_online_mask ((const struct cpumask *)&__cpu_online_mask) 96 #define cpu_present_mask ((const struct cpumask *)&__cpu_present_mask) 97 #define cpu_active_mask ((const struct cpumask *)&__cpu_active_mask) 150 static inline unsigned int cpumask_first(const struct cpumask *srcp) in cpumask_first() [all …]
|
D | pm_opp.h | 156 int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask); 157 int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask); 159 void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask); 352 static inline int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask) in dev_pm_opp_set_sharing_cpus() argument 357 static inline int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) in dev_pm_opp_get_sharing_cpus() argument 366 static inline void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask) in dev_pm_opp_cpumask_remove_table() argument 376 int dev_pm_opp_of_cpumask_add_table(const struct cpumask *cpumask); 377 void dev_pm_opp_of_cpumask_remove_table(const struct cpumask *cpumask); 378 int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask); 383 int dev_pm_opp_of_register_em(struct device *dev, struct cpumask *cpus); [all …]
|
D | stop_machine.h | 37 void stop_machine_yield(const struct cpumask *cpumask); 109 int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); 120 int stop_machine_cpuslocked(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); 123 const struct cpumask *cpus); 127 const struct cpumask *cpus) in stop_machine_cpuslocked() 138 const struct cpumask *cpus) in stop_machine() 144 const struct cpumask *cpus) in stop_machine_from_inactive_cpu()
|
D | interrupt.h | 309 struct cpumask mask; 318 extern int __irq_set_affinity(unsigned int irq, const struct cpumask *cpumask, 329 irq_set_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_set_affinity() argument 331 return __irq_set_affinity(irq, cpumask, false); in irq_set_affinity() 346 irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument 348 return __irq_set_affinity(irq, cpumask, true); in irq_force_affinity() 354 extern int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m); 367 static inline int irq_set_affinity(unsigned int irq, const struct cpumask *m) in irq_set_affinity() 372 static inline int irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument 385 const struct cpumask *m) in irq_set_affinity_hint()
|
D | arch_topology.h | 33 void topology_set_freq_scale(const struct cpumask *cpus, unsigned long cur_freq, 37 bool arch_freq_counters_available(const struct cpumask *cpus); 46 void topology_set_thermal_pressure(const struct cpumask *cpus, 69 const struct cpumask *cpu_coregroup_mask(int cpu);
|
/Linux-v5.10/drivers/opp/ |
D | cpu.c | 108 void _dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask, in _dev_pm_opp_cpumask_remove_table() argument 114 WARN_ON(cpumask_empty(cpumask)); in _dev_pm_opp_cpumask_remove_table() 116 for_each_cpu(cpu, cpumask) { in _dev_pm_opp_cpumask_remove_table() 139 void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask) in dev_pm_opp_cpumask_remove_table() argument 141 _dev_pm_opp_cpumask_remove_table(cpumask, -1); in dev_pm_opp_cpumask_remove_table() 156 const struct cpumask *cpumask) in dev_pm_opp_set_sharing_cpus() argument 167 for_each_cpu(cpu, cpumask) { in dev_pm_opp_set_sharing_cpus() 205 int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) in dev_pm_opp_get_sharing_cpus() argument 220 cpumask_clear(cpumask); in dev_pm_opp_get_sharing_cpus() 225 cpumask_set_cpu(opp_dev->dev->id, cpumask); in dev_pm_opp_get_sharing_cpus() [all …]
|
/Linux-v5.10/kernel/ |
D | padata.c | 67 target_cpu = cpumask_first(pd->cpumask.pcpu); in padata_index_to_cpu() 69 target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); in padata_index_to_cpu() 80 int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); in padata_cpu_hash() 196 if (!cpumask_test_cpu(*cb_cpu, pd->cpumask.cbcpu)) { in padata_do_parallel() 197 if (!cpumask_weight(pd->cpumask.cbcpu)) in padata_do_parallel() 201 cpu_index = *cb_cpu % cpumask_weight(pd->cpumask.cbcpu); in padata_do_parallel() 203 cpu = cpumask_first(pd->cpumask.cbcpu); in padata_do_parallel() 205 cpu = cpumask_next(cpu, pd->cpumask.cbcpu); in padata_do_parallel() 280 pd->cpu = cpumask_next_wrap(cpu, pd->cpumask.pcpu, -1, false); in padata_find_next() 434 cpumask_copy(attrs->cpumask, pinst->cpumask.pcpu); in padata_setup_cpumasks() [all …]
|
D | stop_machine.c | 159 const struct cpumask *active_cpus; 181 notrace void __weak stop_machine_yield(const struct cpumask *cpumask) in stop_machine_yield() argument 192 const struct cpumask *cpumask; in multi_cpu_stop() local 203 cpumask = cpu_online_mask; in multi_cpu_stop() 204 is_active = cpu == cpumask_first(cpumask); in multi_cpu_stop() 206 cpumask = msdata->active_cpus; in multi_cpu_stop() 207 is_active = cpumask_test_cpu(cpu, cpumask); in multi_cpu_stop() 213 stop_machine_yield(cpumask); in multi_cpu_stop() 374 static bool queue_stop_cpus_work(const struct cpumask *cpumask, in queue_stop_cpus_work() argument 390 for_each_cpu(cpu, cpumask) { in queue_stop_cpus_work() [all …]
|
/Linux-v5.10/drivers/powercap/ |
D | idle_inject.c | 73 unsigned long cpumask[]; member 91 for_each_cpu_and(cpu, to_cpumask(ii_dev->cpumask), cpu_online_mask) { in idle_inject_wakeup() 205 cpumask_pr_args(to_cpumask(ii_dev->cpumask))); in idle_inject_start() 234 cpumask_pr_args(to_cpumask(ii_dev->cpumask))); in idle_inject_stop() 253 for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) { in idle_inject_stop() 300 struct idle_inject_device *idle_inject_register(struct cpumask *cpumask) in idle_inject_register() argument 309 cpumask_copy(to_cpumask(ii_dev->cpumask), cpumask); in idle_inject_register() 314 for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) { in idle_inject_register() 327 for_each_cpu(cpu_rb, to_cpumask(ii_dev->cpumask)) { in idle_inject_register() 352 for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) in idle_inject_unregister()
|
/Linux-v5.10/drivers/cpuidle/ |
D | cpuidle-big_little.c | 140 struct cpumask *cpumask; in bl_idle_driver_init() local 143 cpumask = kzalloc(cpumask_size(), GFP_KERNEL); in bl_idle_driver_init() 144 if (!cpumask) in bl_idle_driver_init() 149 cpumask_set_cpu(cpu, cpumask); in bl_idle_driver_init() 151 drv->cpumask = cpumask; in bl_idle_driver_init() 224 kfree(bl_idle_big_driver.cpumask); in bl_idle_init() 226 kfree(bl_idle_little_driver.cpumask); in bl_idle_init()
|
D | dt_idle_states.c | 98 const cpumask_t *cpumask) in idle_state_valid() argument 111 for (cpu = cpumask_next(cpumask_first(cpumask), cpumask); in idle_state_valid() 112 cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { in idle_state_valid() 157 const cpumask_t *cpumask; in dt_init_idle_driver() local 168 cpumask = drv->cpumask ? : cpu_possible_mask; in dt_init_idle_driver() 169 cpu_node = of_cpu_device_node_get(cpumask_first(cpumask)); in dt_init_idle_driver() 187 if (!idle_state_valid(state_node, i, cpumask)) { in dt_init_idle_driver()
|
D | driver.c | 52 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_unset_driver() 72 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_set_driver() 80 for_each_cpu(cpu, drv->cpumask) in __cpuidle_set_driver() 163 if (!drv->cpumask) in __cpuidle_driver_init() 164 drv->cpumask = (struct cpumask *)cpu_possible_mask; in __cpuidle_driver_init() 224 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_register_driver() 243 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_unregister_driver() 362 if (!drv->cpumask) { in cpuidle_driver_state_disabled() 367 for_each_cpu(cpu, drv->cpumask) { in cpuidle_driver_state_disabled()
|
/Linux-v5.10/arch/arc/kernel/ |
D | smp.c | 48 static int __init arc_get_cpu_map(const char *name, struct cpumask *cpumask) in arc_get_cpu_map() argument 57 if (cpulist_parse(buf, cpumask)) in arc_get_cpu_map() 69 struct cpumask cpumask; in arc_init_cpu_possible() local 71 if (arc_get_cpu_map("possible-cpus", &cpumask)) { in arc_init_cpu_possible() 75 cpumask_setall(&cpumask); in arc_init_cpu_possible() 78 if (!cpumask_test_cpu(0, &cpumask)) in arc_init_cpu_possible() 81 init_cpu_possible(&cpumask); in arc_init_cpu_possible() 298 static void ipi_send_msg(const struct cpumask *callmap, enum ipi_msg_type msg) in ipi_send_msg() 313 struct cpumask targets; in smp_send_stop() 324 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|
/Linux-v5.10/arch/riscv/include/asm/ |
D | smp.h | 19 void (*ipi_inject)(const struct cpumask *target); 40 void arch_send_call_function_ipi_mask(struct cpumask *mask); 46 void riscv_cpuid_to_hartid_mask(const struct cpumask *in, struct cpumask *out); 88 static inline void riscv_cpuid_to_hartid_mask(const struct cpumask *in, in riscv_cpuid_to_hartid_mask() 89 struct cpumask *out) in riscv_cpuid_to_hartid_mask()
|
/Linux-v5.10/lib/ |
D | cpumask.c | 18 unsigned int cpumask_next(int n, const struct cpumask *srcp) in cpumask_next() 35 int cpumask_next_and(int n, const struct cpumask *src1p, in cpumask_next_and() 36 const struct cpumask *src2p) in cpumask_next_and() 54 int cpumask_any_but(const struct cpumask *mask, unsigned int cpu) in cpumask_any_but() 78 int cpumask_next_wrap(int n, const struct cpumask *mask, int start, bool wrap) in cpumask_next_wrap() 210 const struct cpumask *mask; in cpumask_local_spread() 252 int cpumask_any_and_distribute(const struct cpumask *src1p, in cpumask_any_and_distribute() 253 const struct cpumask *src2p) in cpumask_any_and_distribute()
|
/Linux-v5.10/arch/x86/kernel/apic/ |
D | local.h | 59 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector); 60 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, int vector); 66 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, int vector); 67 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, int vector); 68 void default_send_IPI_mask_logical(const struct cpumask *mask, int vector);
|
D | ipi.c | 79 void native_send_call_func_ipi(const struct cpumask *mask) in native_send_call_func_ipi() 185 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector) in default_send_IPI_mask_sequence_phys() 203 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, in default_send_IPI_mask_allbutself_phys() 247 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, in default_send_IPI_mask_sequence_logical() 267 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, in default_send_IPI_mask_allbutself_logical() 290 void default_send_IPI_mask_logical(const struct cpumask *cpumask, int vector) in default_send_IPI_mask_logical() argument 292 unsigned long mask = cpumask_bits(cpumask)[0]; in default_send_IPI_mask_logical()
|
D | apic_flat_64.c | 60 static void flat_send_IPI_mask(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask() argument 62 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask() 68 flat_send_IPI_mask_allbutself(const struct cpumask *cpumask, int vector) in flat_send_IPI_mask_allbutself() argument 70 unsigned long mask = cpumask_bits(cpumask)[0]; in flat_send_IPI_mask_allbutself()
|
/Linux-v5.10/drivers/infiniband/hw/hfi1/ |
D | affinity.h | 69 struct cpumask mask; 70 struct cpumask used; 105 struct cpumask general_intr_mask; 106 struct cpumask comp_vect_mask; 112 struct cpumask real_cpu_mask;
|
/Linux-v5.10/arch/powerpc/include/asm/ |
D | smp.h | 119 static inline struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask() 124 static inline struct cpumask *cpu_l2_cache_mask(int cpu) in cpu_l2_cache_mask() 129 static inline struct cpumask *cpu_smallcore_mask(int cpu) in cpu_smallcore_mask() 140 static inline const struct cpumask *cpu_smt_mask(int cpu) in cpu_smt_mask() 192 static inline const struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask() 197 static inline const struct cpumask *cpu_smallcore_mask(int cpu) in cpu_smallcore_mask() 250 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/Linux-v5.10/kernel/time/ |
D | tick-common.c | 206 const struct cpumask *cpumask) in tick_setup_device() argument 258 if (!cpumask_equal(newdev->cpumask, cpumask)) in tick_setup_device() 259 irq_set_affinity(newdev->irq, cpumask); in tick_setup_device() 291 if (!cpumask_test_cpu(cpu, newdev->cpumask)) in tick_check_percpu() 293 if (cpumask_equal(newdev->cpumask, cpumask_of(cpu))) in tick_check_percpu() 299 if (curdev && cpumask_equal(curdev->cpumask, cpumask_of(cpu))) in tick_check_percpu() 321 !cpumask_equal(curdev->cpumask, newdev->cpumask); in tick_check_preferred()
|
/Linux-v5.10/drivers/cpufreq/ |
D | vexpress-spc-cpufreq.c | 283 const struct cpumask *cpumask) in _put_cluster_clk_and_freq_table() argument 295 const struct cpumask *cpumask) in put_cluster_clk_and_freq_table() argument 304 return _put_cluster_clk_and_freq_table(cpu_dev, cpumask); in put_cluster_clk_and_freq_table() 312 _put_cluster_clk_and_freq_table(cdev, cpumask); in put_cluster_clk_and_freq_table() 320 const struct cpumask *cpumask) in _get_cluster_clk_and_freq_table() argument 356 const struct cpumask *cpumask) in get_cluster_clk_and_freq_table() argument 365 ret = _get_cluster_clk_and_freq_table(cpu_dev, cpumask); in get_cluster_clk_and_freq_table() 381 ret = _get_cluster_clk_and_freq_table(cdev, cpumask); in get_cluster_clk_and_freq_table() 404 _put_cluster_clk_and_freq_table(cdev, cpumask); in get_cluster_clk_and_freq_table()
|
/Linux-v5.10/include/trace/events/ |
D | thermal.h | 96 TP_PROTO(const struct cpumask *cpus, unsigned long freq, u32 *load, 102 __bitmask(cpumask, num_possible_cpus()) 110 __assign_bitmask(cpumask, cpumask_bits(cpus), 120 __get_bitmask(cpumask), __entry->freq, 126 TP_PROTO(const struct cpumask *cpus, unsigned int freq, 132 __bitmask(cpumask, num_possible_cpus()) 139 __assign_bitmask(cpumask, cpumask_bits(cpus), 147 __get_bitmask(cpumask), __entry->freq, __entry->cdev_state,
|
/Linux-v5.10/kernel/irq/ |
D | ipi.c | 24 const struct cpumask *dest) in irq_reserve_ipi() 115 int irq_destroy_ipi(unsigned int irq, const struct cpumask *dest) in irq_destroy_ipi() 118 struct cpumask *ipimask = data ? irq_data_get_affinity_mask(data) : NULL; in irq_destroy_ipi() 165 struct cpumask *ipimask = data ? irq_data_get_affinity_mask(data) : NULL; in ipi_get_hwirq() 187 const struct cpumask *dest, unsigned int cpu) in ipi_send_verify() 189 struct cpumask *ipimask = irq_data_get_affinity_mask(data); in ipi_send_verify() 263 int __ipi_send_mask(struct irq_desc *desc, const struct cpumask *dest) in __ipi_send_mask() 328 int ipi_send_mask(unsigned int virq, const struct cpumask *dest) in ipi_send_mask()
|
/Linux-v5.10/arch/ia64/kernel/ |
D | smp.c | 168 send_IPI_mask(const struct cpumask *mask, int op) in send_IPI_mask() 258 cpumask_t cpumask = xcpumask; in smp_flush_tlb_cpumask() local 264 for_each_cpu(cpu, &cpumask) in smp_flush_tlb_cpumask() 268 for_each_cpu(cpu, &cpumask) { in smp_flush_tlb_cpumask() 278 for_each_cpu(cpu, &cpumask) in smp_flush_tlb_cpumask() 323 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|