Home
last modified time | relevance | path

Searched refs:cpumask_weight (Results 1 – 25 of 65) sorted by relevance

123

/Linux-v6.1/lib/
Dcpumask_kunit.c13 "%s contains %sCPUs %*pbl", #m, (cpumask_weight(m) ? "" : "no "), \
19 int mask_weight = cpumask_weight(m); \
29 int mask_weight = cpumask_weight(m); \
43 weight = cpumask_weight(&mask_tmp); \
52 int mask_weight = cpumask_weight(m); \
77 KUNIT_EXPECT_EQ_MSG(test, 0, cpumask_weight(&mask_empty), MASK_MSG(&mask_empty)); in test_cpumask_weight()
78 KUNIT_EXPECT_EQ_MSG(test, nr_cpu_ids, cpumask_weight(cpu_possible_mask), in test_cpumask_weight()
80 KUNIT_EXPECT_EQ_MSG(test, nr_cpumask_bits, cpumask_weight(&mask_all), MASK_MSG(&mask_all)); in test_cpumask_weight()
/Linux-v6.1/arch/x86/include/asm/trace/
Dhyperv.h21 TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
67 TP_fast_assign(__entry->ncpus = cpumask_weight(cpus);
/Linux-v6.1/drivers/infiniband/hw/hfi1/
Daffinity.c105 possible = cpumask_weight(&node_affinity.real_cpu_mask); in init_real_cpu_mask()
106 ht = cpumask_weight(topology_sibling_cpumask( in init_real_cpu_mask()
137 cpumask_weight(topology_sibling_cpumask( in node_affinity_init()
509 if (cpumask_weight(&entry->comp_vect_mask) == 1) { in _dev_comp_vect_cpu_mask_init()
515 cpumask_weight(&entry->comp_vect_mask) / in _dev_comp_vect_cpu_mask_init()
524 cpumask_weight(&entry->comp_vect_mask) % in _dev_comp_vect_cpu_mask_init()
628 possible = cpumask_weight(&entry->def_intr.mask); in hfi1_dev_affinity_init()
974 possible = cpumask_weight(hw_thread_mask); in find_hw_thread_mask()
1019 } else if (current->nr_cpus_allowed < cpumask_weight(&set->mask)) { in hfi1_get_proc_affinity()
/Linux-v6.1/kernel/irq/
Daffinity.c146 ncpus = cpumask_weight(nmsk); in alloc_nodes_vectors()
300 ncpus = cpumask_weight(nmsk); in __irq_build_affinity_masks()
510 set_vecs = cpumask_weight(cpu_possible_mask); in irq_calc_affinity_vectors()
Dipi.c40 nr_irqs = cpumask_weight(dest); in irq_reserve_ipi()
144 nr_irqs = cpumask_weight(dest); in irq_destroy_ipi()
/Linux-v6.1/arch/mips/kernel/
Dcrash.c75 while ((cpumask_weight(&cpus_in_crash) < ncpus) && (--msecs > 0)) { in crash_kexec_prepare_cpus()
/Linux-v6.1/drivers/powercap/
Ddtpm_cpu.c52 nr_cpus = cpumask_weight(&cpus); in set_pd_power_limit()
122 nr_cpus = cpumask_weight(&cpus); in update_pd_power_uw()
/Linux-v6.1/include/linux/
Dcpumask.h638 static inline unsigned int cpumask_weight(const struct cpumask *srcp) in cpumask_weight() function
1024 #define num_possible_cpus() cpumask_weight(cpu_possible_mask)
1025 #define num_present_cpus() cpumask_weight(cpu_present_mask)
1026 #define num_active_cpus() cpumask_weight(cpu_active_mask)
Dtopology.h39 #define nr_cpus_node(node) cpumask_weight(cpumask_of_node(node))
/Linux-v6.1/arch/ia64/include/asm/
Dacpi.h88 low_cpu = cpumask_weight(&early_cpu_possible_map); in per_cpu_scan_finalize()
/Linux-v6.1/drivers/infiniband/sw/siw/
Dsiw_main.c97 if (cpu % cpumask_weight(topology_sibling_cpumask(cpu))) in siw_create_tx_threads()
197 num_cpus = cpumask_weight(tx_cpumask); in siw_get_tx_cpu()
201 num_cpus = cpumask_weight(tx_cpumask); in siw_get_tx_cpu()
/Linux-v6.1/kernel/sched/
Dtopology.c171 if (cpumask_weight(sched_domain_span(sd)) == 1) in sd_degenerate()
374 int i, nr_pd = 0, nr_ps = 0, nr_cpus = cpumask_weight(cpu_map); in build_perf_domains()
683 size = cpumask_weight(sched_domain_span(sd)); in update_top_cache_domain()
981 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span); in init_overlap_sched_group()
1208 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sched_group_span(sg)); in get_group()
1277 sg->group_weight = cpumask_weight(sched_group_span(sg)); in init_sched_groups_capacity()
1556 sd_weight = cpumask_weight(tl->mask(cpu)); in sd_init()
2002 if (cpumask_weight(cpumask_of_node(node)) != 1) in sched_update_numa()
2290 sd->span_weight = cpumask_weight(sched_domain_span(sd)); in build_sched_domains()
/Linux-v6.1/arch/x86/platform/uv/
Duv_nmi.c630 k = n - cpumask_weight(uv_nmi_cpu_mask); in uv_nmi_wait_cpus()
688 cpumask_weight(uv_nmi_cpu_mask), in uv_nmi_wait()
698 cpumask_weight(uv_nmi_cpu_mask), in uv_nmi_wait()
/Linux-v6.1/drivers/thermal/
Dcpufreq_cooling.c272 num_cpus = cpumask_weight(cpufreq_cdev->policy->cpus); in cpufreq_state2power()
354 unsigned int num_cpus = cpumask_weight(cpufreq_cdev->policy->related_cpus); in allocate_idle_time()
/Linux-v6.1/arch/x86/kernel/
Dsmpboot.c643 threads = cpumask_weight(topology_sibling_cpumask(cpu)); in set_cpu_sibling_map()
1592 int threads = cpumask_weight(topology_sibling_cpumask(cpu)); in recompute_smt_state()
1610 if (cpumask_weight(topology_sibling_cpumask(cpu)) == 1) in remove_siblinginfo()
1619 if (cpumask_weight(topology_sibling_cpumask(sibling)) == 1) in remove_siblinginfo()
Dtsc_sync.c343 return (cpumask_weight(topology_core_cpumask(cpu)) > 1) ? 2 : 20; in loop_timeout()
/Linux-v6.1/arch/x86/kernel/cpu/
Dproc.c23 cpumask_weight(topology_core_cpumask(cpu))); in show_cpuinfo_core()
/Linux-v6.1/kernel/
Dstop_machine.c428 cpu_stop_init_done(&done, cpumask_weight(cpumask)); in __stop_cpus()
642 .num_threads = cpumask_weight(smt_mask), in stop_core_cpuslocked()
/Linux-v6.1/drivers/net/wireguard/
Dqueueing.h111 cpu_index = id % cpumask_weight(cpu_online_mask); in wg_cpumask_choose_online()
/Linux-v6.1/arch/loongarch/kernel/
Dmachine_kexec.c219 while ((cpumask_weight(&cpus_in_crash) < ncpus) && timeout--) { in crash_smp_send_stop()
/Linux-v6.1/drivers/net/ethernet/mellanox/mlx5/core/
Dirq_affinity.c61 if (cpumask_weight(req_mask) > 1) in irq_pool_request_irq()
/Linux-v6.1/arch/x86/hyperv/
Dhv_apic.c165 weight = cpumask_weight(mask); in __send_ipi_mask()
/Linux-v6.1/drivers/cpufreq/
Dqcom-cpufreq-hw.c119 for (i = 1; i < cpumask_weight(policy->related_cpus); i++) in qcom_cpufreq_hw_target_index()
160 for (i = 1; i < cpumask_weight(policy->related_cpus); i++) in qcom_cpufreq_hw_fast_switch()
/Linux-v6.1/arch/s390/kernel/
Dprocessor.c294 seq_printf(m, "siblings : %d\n", cpumask_weight(topology_core_cpumask(n))); in show_cpu_topology()
/Linux-v6.1/drivers/firmware/psci/
Dpsci_checker.c93 if (cpumask_weight(offlined_cpus) + 1 == nb_available_cpus) { in down_and_up_cpus()

123