| /Linux-v5.4/include/linux/ |
| D | oom.h | 13 struct zonelist; 31 struct zonelist *zonelist; member
|
| D | mmzone.h | 673 struct zonelist { struct 701 struct zonelist node_zonelists[MAX_ZONELISTS]; 1064 static inline struct zoneref *first_zones_zonelist(struct zonelist *zonelist, in first_zones_zonelist() argument 1068 return next_zones_zonelist(zonelist->_zonerefs, in first_zones_zonelist()
|
| D | swap.h | 352 extern unsigned long try_to_free_pages(struct zonelist *zonelist, int order,
|
| D | gfp.h | 477 static inline struct zonelist *node_zonelist(int nid, gfp_t flags) in node_zonelist()
|
| /Linux-v5.4/mm/ |
| D | mm_init.c | 37 struct zonelist *zonelist; in mminit_verify_zonelist() local 46 zonelist = &pgdat->node_zonelists[listid]; in mminit_verify_zonelist() 57 for_each_zone_zonelist(zone, z, zonelist, zoneid) in mminit_verify_zonelist()
|
| D | vmscan.c | 2950 static void shrink_zones(struct zonelist *zonelist, struct scan_control *sc) in shrink_zones() argument 2970 for_each_zone_zonelist_nodemask(zone, z, zonelist, in shrink_zones() 3066 static unsigned long do_try_to_free_pages(struct zonelist *zonelist, in do_try_to_free_pages() argument 3083 shrink_zones(zonelist, sc); in do_try_to_free_pages() 3100 for_each_zone_zonelist_nodemask(zone, z, zonelist, sc->reclaim_idx, in do_try_to_free_pages() 3177 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument 3215 for_each_zone_zonelist_nodemask(zone, z, zonelist, in throttle_direct_reclaim() 3261 unsigned long try_to_free_pages(struct zonelist *zonelist, int order, in try_to_free_pages() argument 3290 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages() 3296 nr_reclaimed = do_try_to_free_pages(zonelist, &sc); in try_to_free_pages() [all …]
|
| D | internal.h | 114 struct zonelist *zonelist; member
|
| D | oom_kill.c | 271 if (!oc->zonelist) in constrained_alloc() 295 for_each_zone_zonelist_nodemask(zone, z, oc->zonelist, in constrained_alloc() 1124 .zonelist = NULL, in pagefault_out_of_memory()
|
| D | page_alloc.c | 2558 struct zonelist *zonelist = ac->zonelist; in unreserve_highatomic_pageblock() local 2566 for_each_zone_zonelist_nodemask(zone, z, zonelist, ac->high_zoneidx, in unreserve_highatomic_pageblock() 3595 for_next_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in get_page_from_freelist() 3795 .zonelist = ac->zonelist, in __alloc_pages_may_oom() 4046 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in should_compact_retry() 4124 progress = try_to_free_pages(ac->zonelist, order, gfp_mask, in __perform_reclaim() 4175 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, high_zoneidx, in wake_all_kswapds() 4310 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in should_reclaim_retry() 4443 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() 4544 ac->preferred_zoneref = first_zones_zonelist(ac->zonelist, in __alloc_pages_slowpath() [all …]
|
| D | hugetlb.c | 897 struct zonelist *zonelist; in dequeue_huge_page_nodemask() local 902 zonelist = node_zonelist(nid, gfp_mask); in dequeue_huge_page_nodemask() 906 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nmask) { in dequeue_huge_page_nodemask() 1120 struct zonelist *zonelist; in alloc_gigantic_page() local 1124 zonelist = node_zonelist(nid, gfp_mask); in alloc_gigantic_page() 1125 for_each_zone_zonelist_nodemask(zone, z, zonelist, gfp_zone(gfp_mask), nodemask) { in alloc_gigantic_page()
|
| D | mempolicy.c | 1870 struct zonelist *zonelist; in mempolicy_slab_node() local 1872 zonelist = &NODE_DATA(node)->node_zonelists[ZONELIST_FALLBACK]; in mempolicy_slab_node() 1873 z = first_zones_zonelist(zonelist, highest_zoneidx, in mempolicy_slab_node()
|
| D | slab.c | 3105 struct zonelist *zonelist; in fallback_alloc() local 3119 zonelist = node_zonelist(mempolicy_slab_node(), flags); in fallback_alloc() 3126 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in fallback_alloc()
|
| D | compaction.c | 2049 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in compaction_zonelist_suitable() 2358 for_each_zone_zonelist_nodemask(zone, z, ac->zonelist, ac->high_zoneidx, in try_to_compact_pages()
|
| D | slub.c | 1888 struct zonelist *zonelist; in get_any_partial() local 1919 zonelist = node_zonelist(mempolicy_slab_node(), flags); in get_any_partial() 1920 for_each_zone_zonelist(zone, z, zonelist, high_zoneidx) { in get_any_partial()
|
| D | memcontrol.c | 1579 .zonelist = NULL, in mem_cgroup_out_of_memory()
|
| /Linux-v5.4/Documentation/vm/ |
| D | numa.rst | 76 an ordered "zonelist". A zonelist specifies the zones/nodes to visit when a 86 a default Node ordered zonelist. This means it tries to fallback to other zones 91 Linux will attempt to allocate from the first node in the appropriate zonelist 94 nodes' zones in the selected zonelist looking for the first zone in the list 122 zonelist--will not be the node itself. Rather, it will be the node that the
|
| /Linux-v5.4/drivers/tty/ |
| D | sysrq.c | 365 .zonelist = node_zonelist(first_memory_node, gfp_mask), in moom_callback()
|
| /Linux-v5.4/Documentation/admin-guide/sysctl/ |
| D | vm.rst | 597 In non-NUMA case, a zonelist for GFP_KERNEL is ordered as following. 603 Assume 2 node NUMA and below is zonelist of Node(0)'s GFP_KERNEL::
|
| /Linux-v5.4/Documentation/admin-guide/mm/ |
| D | numa_memory_policy.rst | 236 node zonelist.
|
| /Linux-v5.4/Documentation/admin-guide/ |
| D | kernel-parameters.txt | 3219 numa_zonelist_order= [KNL, BOOT] Select zonelist order for NUMA.
|