/Linux-v4.19/include/linux/ |
D | oom.h | 13 struct zonelist; 24 struct zonelist *zonelist; member
|
D | mmzone.h | 606 struct zonelist { struct 626 struct zonelist node_zonelists[MAX_ZONELISTS]; argument 1005 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1009 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist()
|
D | swap.h | 345 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
D | gfp.h | 454 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
/Linux-v4.19/mm/ |
D | mm_init.c | 36 struct zonelist *zonelist; in mminit_verify_zonelist() local 45 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 56 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
D | vmscan.c | 2899 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 2919 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 3019 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 3036 shrink_zones(zonelist, sc); in do_try_to_free_pages() 3053 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages() 3130 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 3168 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 3214 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 3243 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages() 3251 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() [all …]
|
D | internal.h | 118 struct zonelist *zonelist; member
|
D | oom_kill.c | 277 if (!oc->zonelist) in constrained_alloc() 301 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc() 1133 .zonelist = NULL, in pagefault_out_of_memory()
|
D | page_alloc.c | 2285 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 2293 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->high_zoneidx, in unreserve_highatomic_pageblock() 3261 for_next_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in get_page_from_freelist() 3455 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 3690 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in should_compact_retry() 3769 progress = try_to_free_pages(ac->zonelist, order, gfp_mask, in __perform_reclaim() 3820 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, high_zoneidx, in wake_all_kswapds() 3951 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in should_reclaim_retry() 4096 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4173 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() [all …]
|
D | hugetlb.c | 887 struct zonelist *zonelist; in dequeue_huge_page_nodemask() local 892 zonelist = node_zonelist(nid, gfp_mask); in dequeue_huge_page_nodemask() 896 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_huge_page_nodemask() 1110 struct zonelist *zonelist; in alloc_gigantic_page() local 1114 zonelist = node_zonelist(nid, gfp_mask); in alloc_gigantic_page() 1115 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nodemask) { in alloc_gigantic_page()
|
D | mempolicy.c | 1782 struct zonelist *zonelist; in mempolicy_slab_node() local 1784 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 1785 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
D | compaction.c | 1514 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in compaction_zonelist_suitable() 1764 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in try_to_compact_pages()
|
D | slab.c | 3178 struct zonelist *zonelist; in fallback_alloc() local 3192 zonelist = node_zonelist(mempolicy_slab_node(), flags); in fallback_alloc() 3199 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in fallback_alloc()
|
D | slub.c | 1841 struct zonelist *zonelist; in get_any_partial() local 1872 zonelist = node_zonelist(mempolicy_slab_node(), flags); in get_any_partial() 1873 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in get_any_partial()
|
D | memcontrol.c | 1377 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
/Linux-v4.19/Documentation/vm/ |
D | numa.rst | 76 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 86 a default Node ordered zonelist. This means it tries to fallback to other zones 91 Linux will attempt to allocate from the first node in the appropriate zonelist 94 nodes' zones in the selected zonelist looking for the first zone in the list 122 zonelist--will not be the node itself. Rather, it will be the node that the
|
/Linux-v4.19/drivers/tty/ |
D | sysrq.c | 372 .zonelist = node_zonelist(first_memory_node, gfp_mask), in moom_callback()
|
/Linux-v4.19/Documentation/sysctl/ |
D | vm.txt | 580 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 586 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL
|
/Linux-v4.19/Documentation/admin-guide/mm/ |
D | numa_memory_policy.rst | 236 node zonelist.
|
/Linux-v4.19/Documentation/admin-guide/ |
D | kernel-parameters.txt | 2997 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|