/Linux-v5.10/include/linux/ |
D | cpuset.h | 65 int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask); 152 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument 159 current->mems_allowed = nodemask; in set_mems_allowed() 202 static inline int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask) in cpuset_nodemask_valid_mems_allowed() argument 269 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument
|
D | nodemask.h | 152 #define node_isset(node, nodemask) test_bit((node), (nodemask).bits) argument 154 #define node_test_and_set(node, nodemask) \ argument 155 __node_test_and_set((node), &(nodemask)) 231 #define nodes_full(nodemask) __nodes_full(&(nodemask), MAX_NUMNODES) argument 237 #define nodes_weight(nodemask) __nodes_weight(&(nodemask), MAX_NUMNODES) argument
|
D | mempolicy.h | 151 struct mempolicy **mpol, nodemask_t **nodemask); 255 struct mempolicy **mpol, nodemask_t **nodemask) in huge_node() argument 258 *nodemask = NULL; in huge_node()
|
D | mmzone.h | 1117 #define for_each_zone_zonelist_nodemask(zone, z, zlist, highidx, nodemask) \ argument 1118 for (z = first_zones_zonelist(zlist, highidx, nodemask), zone = zonelist_zone(z); \ 1120 z = next_zones_zonelist(++z, highidx, nodemask), \ 1123 #define for_next_zone_zonelist_nodemask(zone, z, highidx, nodemask) \ argument 1126 z = next_zones_zonelist(++z, highidx, nodemask), \
|
D | oom.h | 34 nodemask_t *nodemask; member
|
D | gfp.h | 506 nodemask_t *nodemask); 631 int nid, nodemask_t *nodemask);
|
D | mm.h | 1621 extern void show_free_areas(unsigned int flags, nodemask_t *nodemask); 2448 extern void show_mem(unsigned int flags, nodemask_t *nodemask); 2457 void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...);
|
/Linux-v5.10/lib/ |
D | show_mem.c | 11 void show_mem(unsigned int filter, nodemask_t *nodemask) in show_mem() argument 17 show_free_areas(filter, nodemask); in show_mem()
|
D | Makefile | 39 nmi_backtrace.o nodemask.o win_minmax.o memcat_p.o
|
/Linux-v5.10/mm/ |
D | oom_kill.c | 93 const nodemask_t *mask = oc->nodemask; in oom_cpuset_eligible() 285 if (oc->nodemask && in constrained_alloc() 286 !nodes_subset(node_states[N_MEMORY], *oc->nodemask)) { in constrained_alloc() 288 for_each_node_mask(nid, *oc->nodemask) in constrained_alloc() 295 highest_zoneidx, oc->nodemask) in constrained_alloc() 446 nodemask_pr_args(oc->nodemask)); in dump_oom_summary() 465 show_mem(SHOW_MEM_FILTER_NODES, oc->nodemask); in dump_header() 1088 oc->nodemask = NULL; in out_of_memory() 1129 .nodemask = NULL, in pagefault_out_of_memory()
|
D | page_alloc.c | 2688 ac->nodemask) { in unreserve_highatomic_pageblock() 3804 ac->nodemask) { in get_page_from_freelist() 3937 static void warn_alloc_show_mem(gfp_t gfp_mask, nodemask_t *nodemask) in warn_alloc_show_mem() argument 3953 show_mem(filter, nodemask); in warn_alloc_show_mem() 3956 void warn_alloc(gfp_t gfp_mask, nodemask_t *nodemask, const char *fmt, ...) in warn_alloc() argument 3970 nodemask_pr_args(nodemask)); in warn_alloc() 3976 warn_alloc_show_mem(gfp_mask, nodemask); in warn_alloc() 4005 .nodemask = ac->nodemask, in __alloc_pages_may_oom() 4254 ac->highest_zoneidx, ac->nodemask) { in should_compact_retry() 4331 ac->nodemask); in __perform_reclaim() [all …]
|
D | vmscan.c | 74 nodemask_t *nodemask; member 2920 sc->reclaim_idx, sc->nodemask) { in shrink_zones() 3047 sc->nodemask) { in do_try_to_free_pages() 3150 nodemask_t *nodemask) in throttle_direct_reclaim() argument 3188 gfp_zone(gfp_mask), nodemask) { in throttle_direct_reclaim() 3234 gfp_t gfp_mask, nodemask_t *nodemask) in try_to_free_pages() argument 3242 .nodemask = nodemask, in try_to_free_pages() 3262 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages()
|
D | hugetlb.c | 1099 nodemask_t *nodemask; in dequeue_huge_page_vma() local 1116 nid = huge_node(vma, address, gfp_mask, &mpol, &nodemask); in dequeue_huge_page_vma() 1117 page = dequeue_huge_page_nodemask(h, gfp_mask, nid, nodemask); in dequeue_huge_page_vma() 1239 int nid, nodemask_t *nodemask) in alloc_gigantic_page() argument 1258 for_each_node_mask(node, *nodemask) { in alloc_gigantic_page() 1271 return alloc_contig_pages(nr_pages, gfp_mask, nid, nodemask); in alloc_gigantic_page() 1278 int nid, nodemask_t *nodemask) in alloc_gigantic_page() argument 1286 int nid, nodemask_t *nodemask) in alloc_gigantic_page() argument 1897 nodemask_t *nodemask; in alloc_buddy_huge_page_with_mpol() local 1899 nid = huge_node(vma, addr, gfp_mask, &mpol, &nodemask); in alloc_buddy_huge_page_with_mpol() [all …]
|
D | internal.h | 138 nodemask_t *nodemask; member
|
D | mempolicy.c | 2019 struct mempolicy **mpol, nodemask_t **nodemask) in huge_node() argument 2024 *nodemask = NULL; /* assume !MPOL_BIND */ in huge_node() 2032 *nodemask = &(*mpol)->v.nodes; in huge_node()
|
D | compaction.c | 2172 ac->highest_zoneidx, ac->nodemask) { in compaction_zonelist_suitable() 2487 ac->highest_zoneidx, ac->nodemask) { in try_to_compact_pages()
|
/Linux-v5.10/Documentation/core-api/ |
D | memory-hotplug.rst | 66 - status_change_nid_normal is set node id when N_NORMAL_MEMORY of nodemask 67 is (will be) set/clear, if this is -1, then nodemask status is not changed. 68 - status_change_nid_high is set node id when N_HIGH_MEMORY of nodemask 69 is (will be) set/clear, if this is -1, then nodemask status is not changed. 70 - status_change_nid is set node id when N_MEMORY of nodemask is (will be) 72 node loses all memory. If this is -1, then nodemask status is not changed.
|
D | printk-formats.rst | 522 bitmap and its derivatives such as cpumask and nodemask 530 For printing bitmap and its derivatives such as cpumask and nodemask,
|
/Linux-v5.10/Documentation/admin-guide/mm/ |
D | numa_memory_policy.rst | 217 is always preferred by passing an empty nodemask with this 218 mode. If an empty nodemask is passed, the policy cannot use 251 This flag specifies that the nodemask passed by 257 nodemask (Bind, Interleave) is remapped to the new set of 270 3 is allowed from the user's nodemask, the "interleave" only 271 occurs over that node. If no nodes from the user's nodemask are 276 MPOL_PREFERRED policies that were created with an empty nodemask 280 This flag specifies that the nodemask passed 282 set of allowed nodes. The kernel stores the user-passed nodemask, 283 and if the allowed nodes changes, then that original nodemask will [all …]
|
/Linux-v5.10/kernel/ |
D | smp.c | 582 const struct cpumask *nodemask; in smp_call_function_any() local 591 nodemask = cpumask_of_node(cpu_to_node(cpu)); in smp_call_function_any() 592 for (cpu = cpumask_first_and(nodemask, mask); cpu < nr_cpu_ids; in smp_call_function_any() 593 cpu = cpumask_next_and(cpu, nodemask, mask)) { in smp_call_function_any()
|
/Linux-v5.10/tools/perf/bench/ |
D | numa.c | 347 unsigned long nodemask; in bind_to_memnode() local 353 BUG_ON(g->p.nr_nodes > (int)sizeof(nodemask)*8); in bind_to_memnode() 354 nodemask = 1L << node; in bind_to_memnode() 356 ret = set_mempolicy(MPOL_BIND, &nodemask, sizeof(nodemask)*8); in bind_to_memnode() 357 dprintf("binding to node %d, mask: %016lx => %d\n", node, nodemask, ret); in bind_to_memnode()
|
/Linux-v5.10/arch/x86/mm/ |
D | numa.c | 329 static void __init numa_nodemask_from_meminfo(nodemask_t *nodemask, in numa_nodemask_from_meminfo() argument 337 node_set(mi->blk[i].nid, *nodemask); in numa_nodemask_from_meminfo()
|
/Linux-v5.10/kernel/irq/ |
D | manage.c | 512 const struct cpumask *nodemask = cpumask_of_node(node); in irq_setup_affinity() local 515 if (cpumask_intersects(&mask, nodemask)) in irq_setup_affinity() 516 cpumask_and(&mask, &mask, nodemask); in irq_setup_affinity()
|
/Linux-v5.10/drivers/tty/ |
D | sysrq.c | 382 .nodemask = NULL, in moom_callback()
|
/Linux-v5.10/kernel/cgroup/ |
D | cpuset.c | 3361 int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask) in cpuset_nodemask_valid_mems_allowed() argument 3363 return nodes_intersects(*nodemask, current->mems_allowed); in cpuset_nodemask_valid_mems_allowed()
|