/Linux-v5.4/include/linux/ |
D | cpumask.h | 17 typedef struct cpumask { DECLARE_BITMAP(bits, NR_CPUS); } cpumask_t; struct 90 extern struct cpumask __cpu_possible_mask; 91 extern struct cpumask __cpu_online_mask; 92 extern struct cpumask __cpu_present_mask; 93 extern struct cpumask __cpu_active_mask; 94 #define cpu_possible_mask ((const struct cpumask *)&__cpu_possible_mask) 95 #define cpu_online_mask ((const struct cpumask *)&__cpu_online_mask) 96 #define cpu_present_mask ((const struct cpumask *)&__cpu_present_mask) 97 #define cpu_active_mask ((const struct cpumask *)&__cpu_active_mask) 150 static inline unsigned int cpumask_first(const struct cpumask *srcp) in cpumask_first() [all …]
|
D | stop_machine.h | 35 int stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg); 36 int try_stop_cpus(const struct cpumask *cpumask, cpu_stop_fn_t fn, void *arg); 39 void stop_machine_yield(const struct cpumask *cpumask); 85 static inline int stop_cpus(const struct cpumask *cpumask, in stop_cpus() argument 88 if (cpumask_test_cpu(raw_smp_processor_id(), cpumask)) in stop_cpus() 93 static inline int try_stop_cpus(const struct cpumask *cpumask, in try_stop_cpus() argument 96 return stop_cpus(cpumask, fn, arg); in try_stop_cpus() 125 int stop_machine(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); 136 int stop_machine_cpuslocked(cpu_stop_fn_t fn, void *data, const struct cpumask *cpus); 139 const struct cpumask *cpus); [all …]
|
D | pm_opp.h | 137 int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask); 138 int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask); 140 void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask); 320 static inline int dev_pm_opp_set_sharing_cpus(struct device *cpu_dev, const struct cpumask *cpumask) in dev_pm_opp_set_sharing_cpus() argument 325 static inline int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) in dev_pm_opp_get_sharing_cpus() argument 334 static inline void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask) in dev_pm_opp_cpumask_remove_table() argument 344 int dev_pm_opp_of_cpumask_add_table(const struct cpumask *cpumask); 345 void dev_pm_opp_of_cpumask_remove_table(const struct cpumask *cpumask); 346 int dev_pm_opp_of_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask); 350 void dev_pm_opp_of_register_em(struct cpumask *cpus); [all …]
|
D | interrupt.h | 294 struct cpumask mask; 303 extern int __irq_set_affinity(unsigned int irq, const struct cpumask *cpumask, 314 irq_set_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_set_affinity() argument 316 return __irq_set_affinity(irq, cpumask, false); in irq_set_affinity() 331 irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument 333 return __irq_set_affinity(irq, cpumask, true); in irq_force_affinity() 339 extern int irq_set_affinity_hint(unsigned int irq, const struct cpumask *m); 352 static inline int irq_set_affinity(unsigned int irq, const struct cpumask *m) in irq_set_affinity() 357 static inline int irq_force_affinity(unsigned int irq, const struct cpumask *cpumask) in irq_force_affinity() argument 370 const struct cpumask *m) in irq_set_affinity_hint()
|
/Linux-v5.4/kernel/ |
D | padata.c | 42 target_cpu = cpumask_first(pd->cpumask.pcpu); in padata_index_to_cpu() 44 target_cpu = cpumask_next(target_cpu, pd->cpumask.pcpu); in padata_index_to_cpu() 55 int cpu_index = seq_nr % cpumask_weight(pd->cpumask.pcpu); in padata_cpu_hash() 116 if (!cpumask_test_cpu(*cb_cpu, pd->cpumask.cbcpu)) { in padata_do_parallel() 117 if (!cpumask_weight(pd->cpumask.cbcpu)) in padata_do_parallel() 121 cpu_index = *cb_cpu % cpumask_weight(pd->cpumask.cbcpu); in padata_do_parallel() 123 cpu = cpumask_first(pd->cpumask.cbcpu); in padata_do_parallel() 125 cpu = cpumask_next(cpu, pd->cpumask.cbcpu); in padata_do_parallel() 204 pd->cpu = cpumask_next_wrap(cpu, pd->cpumask.pcpu, -1, false); in padata_find_next() 345 const struct cpumask *pcpumask, in padata_setup_cpumasks() [all …]
|
D | stop_machine.c | 159 const struct cpumask *active_cpus; 181 void __weak stop_machine_yield(const struct cpumask *cpumask) in stop_machine_yield() argument 192 const struct cpumask *cpumask; in multi_cpu_stop() local 203 cpumask = cpu_online_mask; in multi_cpu_stop() 204 is_active = cpu == cpumask_first(cpumask); in multi_cpu_stop() 206 cpumask = msdata->active_cpus; in multi_cpu_stop() 207 is_active = cpumask_test_cpu(cpu, cpumask); in multi_cpu_stop() 213 stop_machine_yield(cpumask); in multi_cpu_stop() 373 static bool queue_stop_cpus_work(const struct cpumask *cpumask, in queue_stop_cpus_work() argument 389 for_each_cpu(cpu, cpumask) { in queue_stop_cpus_work() [all …]
|
D | smp.c | 33 cpumask_var_t cpumask; member 47 if (!zalloc_cpumask_var_node(&cfd->cpumask, GFP_KERNEL, in smpcfd_prepare_cpu() 52 free_cpumask_var(cfd->cpumask); in smpcfd_prepare_cpu() 57 free_cpumask_var(cfd->cpumask); in smpcfd_prepare_cpu() 69 free_cpumask_var(cfd->cpumask); in smpcfd_dead_cpu() 369 int smp_call_function_any(const struct cpumask *mask, in smp_call_function_any() 373 const struct cpumask *nodemask; in smp_call_function_any() 412 void smp_call_function_many(const struct cpumask *mask, in smp_call_function_many() 457 cpumask_and(cfd->cpumask, mask, cpu_online_mask); in smp_call_function_many() 458 __cpumask_clear_cpu(this_cpu, cfd->cpumask); in smp_call_function_many() [all …]
|
/Linux-v5.4/drivers/opp/ |
D | cpu.c | 108 void _dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask, in _dev_pm_opp_cpumask_remove_table() argument 114 WARN_ON(cpumask_empty(cpumask)); in _dev_pm_opp_cpumask_remove_table() 116 for_each_cpu(cpu, cpumask) { in _dev_pm_opp_cpumask_remove_table() 139 void dev_pm_opp_cpumask_remove_table(const struct cpumask *cpumask) in dev_pm_opp_cpumask_remove_table() argument 141 _dev_pm_opp_cpumask_remove_table(cpumask, -1); in dev_pm_opp_cpumask_remove_table() 156 const struct cpumask *cpumask) in dev_pm_opp_set_sharing_cpus() argument 167 for_each_cpu(cpu, cpumask) { in dev_pm_opp_set_sharing_cpus() 205 int dev_pm_opp_get_sharing_cpus(struct device *cpu_dev, struct cpumask *cpumask) in dev_pm_opp_get_sharing_cpus() argument 220 cpumask_clear(cpumask); in dev_pm_opp_get_sharing_cpus() 225 cpumask_set_cpu(opp_dev->dev->id, cpumask); in dev_pm_opp_get_sharing_cpus() [all …]
|
/Linux-v5.4/arch/x86/include/asm/uv/ |
D | uv.h | 9 struct cpumask; 27 extern const struct cpumask *uv_flush_tlb_others(const struct cpumask *cpumask, 38 static inline const struct cpumask * 39 uv_flush_tlb_others(const struct cpumask *cpumask, in uv_flush_tlb_others() argument 41 { return cpumask; } in uv_flush_tlb_others()
|
/Linux-v5.4/arch/arc/plat-eznps/ |
D | smp.c | 21 static int __init eznps_get_map(const char *name, struct cpumask *cpumask) in eznps_get_map() argument 30 cpulist_parse(buf, cpumask); in eznps_get_map() 38 struct cpumask cpumask; in eznps_init_cpumasks() local 40 if (eznps_get_map("present-cpus", &cpumask)) { in eznps_init_cpumasks() 44 init_cpu_present(&cpumask); in eznps_init_cpumasks() 46 if (eznps_get_map("possible-cpus", &cpumask)) { in eznps_init_cpumasks() 50 init_cpu_possible(&cpumask); in eznps_init_cpumasks()
|
/Linux-v5.4/drivers/powercap/ |
D | idle_inject.c | 70 unsigned long int cpumask[0]; member 88 for_each_cpu_and(cpu, to_cpumask(ii_dev->cpumask), cpu_online_mask) { in idle_inject_wakeup() 191 cpumask_pr_args(to_cpumask(ii_dev->cpumask))); in idle_inject_start() 220 cpumask_pr_args(to_cpumask(ii_dev->cpumask))); in idle_inject_stop() 239 for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) { in idle_inject_stop() 288 struct idle_inject_device *idle_inject_register(struct cpumask *cpumask) in idle_inject_register() argument 297 cpumask_copy(to_cpumask(ii_dev->cpumask), cpumask); in idle_inject_register() 301 for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) { in idle_inject_register() 314 for_each_cpu(cpu_rb, to_cpumask(ii_dev->cpumask)) { in idle_inject_register() 339 for_each_cpu(cpu, to_cpumask(ii_dev->cpumask)) in idle_inject_unregister()
|
/Linux-v5.4/drivers/cpuidle/ |
D | cpuidle-big_little.c | 140 struct cpumask *cpumask; in bl_idle_driver_init() local 143 cpumask = kzalloc(cpumask_size(), GFP_KERNEL); in bl_idle_driver_init() 144 if (!cpumask) in bl_idle_driver_init() 149 cpumask_set_cpu(cpu, cpumask); in bl_idle_driver_init() 151 drv->cpumask = cpumask; in bl_idle_driver_init() 224 kfree(bl_idle_big_driver.cpumask); in bl_idle_init() 226 kfree(bl_idle_little_driver.cpumask); in bl_idle_init()
|
D | dt_idle_states.c | 98 const cpumask_t *cpumask) in idle_state_valid() argument 111 for (cpu = cpumask_next(cpumask_first(cpumask), cpumask); in idle_state_valid() 112 cpu < nr_cpu_ids; cpu = cpumask_next(cpu, cpumask)) { in idle_state_valid() 158 const cpumask_t *cpumask; in dt_init_idle_driver() local 169 cpumask = drv->cpumask ? : cpu_possible_mask; in dt_init_idle_driver() 170 cpu_node = of_cpu_device_node_get(cpumask_first(cpumask)); in dt_init_idle_driver() 188 if (!idle_state_valid(state_node, i, cpumask)) { in dt_init_idle_driver()
|
D | driver.c | 52 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_unset_driver() 74 for_each_cpu(cpu, drv->cpumask) { in __cpuidle_set_driver() 166 if (!drv->cpumask) in __cpuidle_driver_init() 167 drv->cpumask = (struct cpumask *)cpu_possible_mask; in __cpuidle_driver_init() 216 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_register_driver() 238 on_each_cpu_mask(drv->cpumask, cpuidle_setup_broadcast_timer, in __cpuidle_unregister_driver()
|
/Linux-v5.4/arch/arc/kernel/ |
D | smp.c | 48 static int __init arc_get_cpu_map(const char *name, struct cpumask *cpumask) in arc_get_cpu_map() argument 57 if (cpulist_parse(buf, cpumask)) in arc_get_cpu_map() 69 struct cpumask cpumask; in arc_init_cpu_possible() local 71 if (arc_get_cpu_map("possible-cpus", &cpumask)) { in arc_init_cpu_possible() 75 cpumask_setall(&cpumask); in arc_init_cpu_possible() 78 if (!cpumask_test_cpu(0, &cpumask)) in arc_init_cpu_possible() 81 init_cpu_possible(&cpumask); in arc_init_cpu_possible() 298 static void ipi_send_msg(const struct cpumask *callmap, enum ipi_msg_type msg) in ipi_send_msg() 313 struct cpumask targets; in smp_send_stop() 324 void arch_send_call_function_ipi_mask(const struct cpumask *mask) in arch_send_call_function_ipi_mask()
|
/Linux-v5.4/drivers/cpufreq/ |
D | arm_big_little.h | 33 int (*init_opp_table)(const struct cpumask *cpumask); 37 void (*free_opp_table)(const struct cpumask *cpumask);
|
D | arm_big_little.c | 310 const struct cpumask *cpumask) in _put_cluster_clk_and_freq_table() argument 320 arm_bL_ops->free_opp_table(cpumask); in _put_cluster_clk_and_freq_table() 325 const struct cpumask *cpumask) in put_cluster_clk_and_freq_table() argument 334 return _put_cluster_clk_and_freq_table(cpu_dev, cpumask); in put_cluster_clk_and_freq_table() 343 _put_cluster_clk_and_freq_table(cdev, cpumask); in put_cluster_clk_and_freq_table() 351 const struct cpumask *cpumask) in _get_cluster_clk_and_freq_table() argument 359 ret = arm_bL_ops->init_opp_table(cpumask); in _get_cluster_clk_and_freq_table() 388 arm_bL_ops->free_opp_table(cpumask); in _get_cluster_clk_and_freq_table() 396 const struct cpumask *cpumask) in get_cluster_clk_and_freq_table() argument 405 ret = _get_cluster_clk_and_freq_table(cpu_dev, cpumask); in get_cluster_clk_and_freq_table() [all …]
|
/Linux-v5.4/arch/x86/kernel/apic/ |
D | local.h | 58 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector); 59 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, int vector); 65 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, int vector); 66 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, int vector); 67 void default_send_IPI_mask_logical(const struct cpumask *mask, int vector);
|
D | ipi.c | 78 void native_send_call_func_ipi(const struct cpumask *mask) in native_send_call_func_ipi() 184 void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int vector) in default_send_IPI_mask_sequence_phys() 202 void default_send_IPI_mask_allbutself_phys(const struct cpumask *mask, in default_send_IPI_mask_allbutself_phys() 246 void default_send_IPI_mask_sequence_logical(const struct cpumask *mask, in default_send_IPI_mask_sequence_logical() 266 void default_send_IPI_mask_allbutself_logical(const struct cpumask *mask, in default_send_IPI_mask_allbutself_logical() 289 void default_send_IPI_mask_logical(const struct cpumask *cpumask, int vector) in default_send_IPI_mask_logical() argument 291 unsigned long mask = cpumask_bits(cpumask)[0]; in default_send_IPI_mask_logical()
|
/Linux-v5.4/drivers/infiniband/hw/hfi1/ |
D | affinity.h | 68 struct cpumask mask; 69 struct cpumask used; 104 struct cpumask general_intr_mask; 105 struct cpumask comp_vect_mask; 111 struct cpumask real_cpu_mask;
|
/Linux-v5.4/arch/arm64/include/asm/ |
D | smp.h | 72 extern void set_smp_cross_call(void (*)(const struct cpumask *, unsigned int)); 74 extern void (*__smp_cross_call)(const struct cpumask *, unsigned int); 98 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask); 101 extern void arch_send_wakeup_ipi_mask(const struct cpumask *mask); 103 static inline void arch_send_wakeup_ipi_mask(const struct cpumask *mask) in arch_send_wakeup_ipi_mask()
|
/Linux-v5.4/arch/powerpc/include/asm/ |
D | smp.h | 116 static inline struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask() 121 static inline struct cpumask *cpu_core_mask(int cpu) in cpu_core_mask() 126 static inline struct cpumask *cpu_l2_cache_mask(int cpu) in cpu_l2_cache_mask() 131 static inline struct cpumask *cpu_smallcore_mask(int cpu) in cpu_smallcore_mask() 181 static inline const struct cpumask *cpu_sibling_mask(int cpu) in cpu_sibling_mask() 186 static inline const struct cpumask *cpu_smallcore_mask(int cpu) in cpu_smallcore_mask() 239 extern void arch_send_call_function_ipi_mask(const struct cpumask *mask);
|
/Linux-v5.4/kernel/time/ |
D | tick-common.c | 203 const struct cpumask *cpumask) in tick_setup_device() argument 255 if (!cpumask_equal(newdev->cpumask, cpumask)) in tick_setup_device() 256 irq_set_affinity(newdev->irq, cpumask); in tick_setup_device() 288 if (!cpumask_test_cpu(cpu, newdev->cpumask)) in tick_check_percpu() 290 if (cpumask_equal(newdev->cpumask, cpumask_of(cpu))) in tick_check_percpu() 296 if (curdev && cpumask_equal(curdev->cpumask, cpumask_of(cpu))) in tick_check_percpu() 318 !cpumask_equal(curdev->cpumask, newdev->cpumask); in tick_check_preferred()
|
/Linux-v5.4/include/trace/events/ |
D | thermal.h | 96 TP_PROTO(const struct cpumask *cpus, unsigned long freq, u32 *load, 102 __bitmask(cpumask, num_possible_cpus()) 110 __assign_bitmask(cpumask, cpumask_bits(cpus), 120 __get_bitmask(cpumask), __entry->freq, 126 TP_PROTO(const struct cpumask *cpus, unsigned int freq, 132 __bitmask(cpumask, num_possible_cpus()) 139 __assign_bitmask(cpumask, cpumask_bits(cpus), 147 __get_bitmask(cpumask), __entry->freq, __entry->cdev_state,
|
/Linux-v5.4/kernel/irq/ |
D | ipi.c | 24 const struct cpumask *dest) in irq_reserve_ipi() 115 int irq_destroy_ipi(unsigned int irq, const struct cpumask *dest) in irq_destroy_ipi() 118 struct cpumask *ipimask = data ? irq_data_get_affinity_mask(data) : NULL; in irq_destroy_ipi() 165 struct cpumask *ipimask = data ? irq_data_get_affinity_mask(data) : NULL; in ipi_get_hwirq() 187 const struct cpumask *dest, unsigned int cpu) in ipi_send_verify() 189 struct cpumask *ipimask = irq_data_get_affinity_mask(data); in ipi_send_verify() 263 int __ipi_send_mask(struct irq_desc *desc, const struct cpumask *dest) in __ipi_send_mask() 328 int ipi_send_mask(unsigned int virq, const struct cpumask *dest) in ipi_send_mask()
|