/Linux-v5.15/include/linux/ |
D | oom.h | 13 struct zonelist; 31 struct zonelist *zonelist; member
|
D | mmzone.h | 773 struct zonelist { struct 813 struct zonelist node_zonelists[MAX_ZONELISTS]; 1179 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1183 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist()
|
D | swap.h | 379 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
D | gfp.h | 512 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
/Linux-v5.15/mm/ |
D | mm_init.c | 34 struct zonelist *zonelist; in mminit_verify_zonelist() local 43 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 54 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
D | vmscan.c | 3267 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 3287 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 3380 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 3397 shrink_zones(zonelist, sc); in do_try_to_free_pages() 3414 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages() 3517 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 3555 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 3597 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 3626 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages() 3632 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() [all …]
|
D | internal.h | 141 struct zonelist *zonelist; member
|
D | oom_kill.c | 273 if (!oc->zonelist) in constrained_alloc() 297 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc() 1130 .zonelist = NULL, in pagefault_out_of_memory()
|
D | page_alloc.c | 2813 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 2821 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->highest_zoneidx, in unreserve_highatomic_pageblock() 4255 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 4509 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_compact_retry() 4594 progress = try_to_free_pages(ac->zonelist, order, gfp_mask, in __perform_reclaim() 4645 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, highest_zoneidx, in wake_all_kswapds() 4781 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in should_reclaim_retry() 4914 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 5001 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 5137 ac->zonelist = node_zonelist(preferred_nid, gfp_mask); in prepare_alloc_pages() [all …]
|
D | mempolicy.c | 1851 struct zonelist *zonelist; in mempolicy_slab_node() local 1853 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 1854 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
D | slab.c | 3099 struct zonelist *zonelist; in fallback_alloc() local 3113 zonelist = node_zonelist(mempolicy_slab_node(), flags); in fallback_alloc() 3120 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in fallback_alloc()
|
D | compaction.c | 2246 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in compaction_zonelist_suitable() 2569 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, in try_to_compact_pages()
|
D | hugetlb.c | 1111 struct zonelist *zonelist; in dequeue_huge_page_nodemask() local 1116 zonelist = node_zonelist(nid, gfp_mask); in dequeue_huge_page_nodemask() 1120 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_huge_page_nodemask()
|
D | slub.c | 2158 struct zonelist *zonelist; in get_any_partial() local 2189 zonelist = node_zonelist(mempolicy_slab_node(), flags); in get_any_partial() 2190 for_each_zone_zonelist(zone, z, zonelist, highest_zoneidx) { in get_any_partial()
|
D | memcontrol.c | 1560 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
/Linux-v5.15/Documentation/vm/ |
D | numa.rst | 76 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 86 a default Node ordered zonelist. This means it tries to fallback to other zones 91 Linux will attempt to allocate from the first node in the appropriate zonelist 94 nodes' zones in the selected zonelist looking for the first zone in the list 122 zonelist--will not be the node itself. Rather, it will be the node that the
|
/Linux-v5.15/drivers/tty/ |
D | sysrq.c | 382 .zonelist = node_zonelist(first_memory_node, gfp_mask), in moom_callback()
|
/Linux-v5.15/Documentation/admin-guide/sysctl/ |
D | vm.rst | 610 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 616 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL::
|
/Linux-v5.15/Documentation/admin-guide/mm/ |
D | numa_memory_policy.rst | 236 node zonelist.
|
/Linux-v5.15/Documentation/admin-guide/ |
D | kernel-parameters.txt | 3614 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|