Home
last modified time | relevance | path

Searched refs:nr_node_ids (Results 1 – 25 of 28) sorted by relevance

12

/Linux-v5.4/kernel/irq/
Daffinity.c48 masks = kcalloc(nr_node_ids, sizeof(cpumask_var_t), GFP_KERNEL); in alloc_node_to_cpumask()
52 for (node = 0; node < nr_node_ids; node++) { in alloc_node_to_cpumask()
70 for (node = 0; node < nr_node_ids; node++) in free_node_to_cpumask()
137 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_vectors()
156 sort(node_vectors, nr_node_ids, sizeof(node_vectors[0]), in alloc_nodes_vectors()
227 for (n = 0; n < nr_node_ids; n++) { in alloc_nodes_vectors()
280 node_vectors = kcalloc(nr_node_ids, in __irq_build_affinity_masks()
290 for (i = 0; i < nr_node_ids; i++) { in __irq_build_affinity_masks()
/Linux-v5.4/arch/arm64/mm/
Dnuma.c49 if (WARN_ON(node >= nr_node_ids)) in cpumask_of_node()
102 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map()
106 for (node = 0; node < nr_node_ids; node++) { in setup_node_to_cpumask_map()
112 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map()
278 size = nr_node_ids * nr_node_ids * sizeof(numa_distance[0]); in numa_alloc_distance()
287 numa_distance_cnt = nr_node_ids; in numa_alloc_distance()
/Linux-v5.4/arch/x86/mm/
Dnuma.c119 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map()
123 for (node = 0; node < nr_node_ids; node++) in setup_node_to_cpumask_map()
127 pr_debug("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map()
864 if ((unsigned)node >= nr_node_ids) { in cpumask_of_node()
867 node, nr_node_ids); in cpumask_of_node()
/Linux-v5.4/kernel/sched/
Dtopology.c195 if (nr_node_ids == 1) in sd_parent_degenerate()
1479 for (i = 0; i < nr_node_ids; i++) { in sched_numa_warn()
1481 for (j = 0; j < nr_node_ids; j++) in sched_numa_warn()
1562 sched_domains_numa_distance = kzalloc(sizeof(int) * (nr_node_ids + 1), GFP_KERNEL); in sched_init_numa()
1578 for (i = 0; i < nr_node_ids; i++) { in sched_init_numa()
1579 for (j = 0; j < nr_node_ids; j++) { in sched_init_numa()
1580 for (k = 0; k < nr_node_ids; k++) { in sched_init_numa()
1641 kzalloc(nr_node_ids * sizeof(void *), GFP_KERNEL); in sched_init_numa()
1645 for (j = 0; j < nr_node_ids; j++) { in sched_init_numa()
1711 for (j = 0; j < nr_node_ids; j++) { in sched_domains_numa_masks_set()
[all …]
Dfair.c1233 return NR_NUMA_HINT_FAULT_TYPES * (s * nr_node_ids + nid) + priv; in task_faults_idx()
2255 4*nr_node_ids*sizeof(unsigned long); in task_numa_group()
2268 nr_node_ids; in task_numa_group()
2270 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) in task_numa_group()
2328 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) { in task_numa_group()
2371 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) in task_numa_free()
2386 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) in task_numa_free()
2413 NR_NUMA_HINT_FAULT_BUCKETS * nr_node_ids; in task_numa_fault()
/Linux-v5.4/arch/x86/kernel/
Dsetup_percpu.c175 NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids); in setup_per_cpu_areas()
/Linux-v5.4/include/linux/
Dnodemask.h447 extern unsigned int nr_node_ids;
488 #define nr_node_ids 1U macro
/Linux-v5.4/drivers/hv/
Dhv.c89 hv_context.hv_numa_map = kcalloc(nr_node_ids, sizeof(struct cpumask), in hv_synic_alloc()
Dchannel_mgmt.c687 if (next_node == nr_node_ids) { in init_vp_index()
/Linux-v5.4/arch/powerpc/mm/
Dnuma.c74 if (nr_node_ids == MAX_NUMNODES) in setup_node_to_cpumask_map()
82 dbg("Node to cpumask map for %u nodes\n", nr_node_ids); in setup_node_to_cpumask_map()
/Linux-v5.4/mm/
Dksm.c2974 buf = kcalloc(nr_node_ids + nr_node_ids, sizeof(*buf), in merge_across_nodes_store()
2981 root_unstable_tree = buf + nr_node_ids; in merge_across_nodes_store()
2988 ksm_nr_node_ids = knob ? 1 : nr_node_ids; in merge_across_nodes_store()
Dslab.h638 for (__node = 0; __node < nr_node_ids; __node++) \
Dlist_lru.c614 lru->node = kcalloc(nr_node_ids, sizeof(*lru->node), GFP_KERNEL); in __list_lru_init()
Dmempolicy.c1392 unsigned int nbytes = BITS_TO_LONGS(nr_node_ids) * sizeof(long); in copy_nodes_to_user()
1568 if (nmask != NULL && maxnode < nr_node_ids) in kernel_get_mempolicy()
1604 nr_bits = min_t(unsigned long, maxnode-1, nr_node_ids); in COMPAT_SYSCALL_DEFINE5()
Dslub.c4244 nr_node_ids * sizeof(struct kmem_cache_node *), in kmem_cache_init()
4263 nr_cpu_ids, nr_node_ids); in kmem_cache_init()
4795 nodes = kcalloc(nr_node_ids, sizeof(unsigned long), GFP_KERNEL); in show_slab_objects()
4884 for (node = 0; node < nr_node_ids; node++) in show_slab_objects()
Dvmalloc.c3450 memset(counters, 0, nr_node_ids * sizeof(unsigned int)); in show_numa_info()
3555 nr_node_ids * sizeof(unsigned int), NULL); in proc_vmalloc_init()
Dswapfile.c2830 p = kvzalloc(struct_size(p, avail_lists, nr_node_ids), GFP_KERNEL); in alloc_swap_info()
3775 swap_avail_heads = kmalloc_array(nr_node_ids, sizeof(struct plist_head), in swapfile_init()
Dslab.c655 alc_ptr = kcalloc_node(nr_node_ids, sizeof(void *), gfp, node); in alloc_alien_cache()
1239 nr_node_ids * sizeof(struct kmem_cache_node *), in kmem_cache_init()
Dcompaction.c2473 if (nid >= 0 && nid < nr_node_ids && node_online(nid)) { in sysfs_compact_node()
Dpage_alloc.c354 unsigned int nr_node_ids __read_mostly = MAX_NUMNODES;
356 EXPORT_SYMBOL(nr_node_ids);
6987 nr_node_ids = highest + 1; in setup_nr_node_ids()
Dmemcontrol.c3757 VM_BUG_ON((unsigned)nid >= nr_node_ids); in mem_cgroup_node_nr_lru_pages()
5053 size += nr_node_ids * sizeof(struct mem_cgroup_per_node *); in mem_cgroup_alloc()
/Linux-v5.4/kernel/
Dworkqueue.c3896 ctx = kzalloc(struct_size(ctx, pwq_tbl, nr_node_ids), GFP_KERNEL); in apply_wqattrs_prepare()
4245 tbl_size = nr_node_ids * sizeof(wq->numa_pwq_tbl[0]); in alloc_workqueue()
5835 tbl = kcalloc(nr_node_ids, sizeof(tbl[0]), GFP_KERNEL); in wq_numa_init()
/Linux-v5.4/drivers/net/ethernet/cavium/thunder/
Dnic_main.c1244 if (nr_node_ids > 1) in nic_num_sqs_en()
/Linux-v5.4/net/sunrpc/
Dsvc.c196 unsigned int maxpools = nr_node_ids; in svc_pool_map_init_pernode()
/Linux-v5.4/drivers/char/
Drandom.c899 pool = kcalloc(nr_node_ids, sizeof(*pool), GFP_KERNEL|__GFP_NOFAIL); in do_numa_crng_init()

12