Lines Matching refs:gfp_mask
133 gfp_t gfp_mask; member
545 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid, in shrink_slab_memcg() argument
565 .gfp_mask = gfp_mask, in shrink_slab_memcg()
620 static unsigned long shrink_slab_memcg(gfp_t gfp_mask, int nid, in shrink_slab_memcg() argument
647 static unsigned long shrink_slab(gfp_t gfp_mask, int nid, in shrink_slab() argument
662 return shrink_slab_memcg(gfp_mask, nid, memcg, priority); in shrink_slab()
669 .gfp_mask = gfp_mask, in shrink_slab()
1115 may_enter_fs = (sc->gfp_mask & __GFP_FS) || in shrink_page_list()
1116 (PageSwapCache(page) && (sc->gfp_mask & __GFP_IO)); in shrink_page_list()
1242 if (!(sc->gfp_mask & __GFP_IO)) in shrink_page_list()
1402 if (!try_to_release_page(page, sc->gfp_mask)) in shrink_page_list()
1499 .gfp_mask = GFP_KERNEL, in reclaim_clean_pages_from_list()
1818 if ((sc->gfp_mask & (__GFP_IO | __GFP_FS)) == (__GFP_IO | __GFP_FS)) in too_many_isolated()
2112 .gfp_mask = GFP_KERNEL, in reclaim_pages()
2656 shrink_slab(sc->gfp_mask, pgdat->node_id, memcg, in shrink_node_memcgs()
2660 vmpressure(sc->gfp_mask, memcg, false, in shrink_node_memcgs()
2779 vmpressure(sc->gfp_mask, sc->target_mem_cgroup, true, in shrink_node()
2913 orig_mask = sc->gfp_mask; in shrink_zones()
2915 sc->gfp_mask |= __GFP_HIGHMEM; in shrink_zones()
2916 sc->reclaim_idx = gfp_zone(sc->gfp_mask); in shrink_zones()
2963 sc->order, sc->gfp_mask, in shrink_zones()
2981 sc->gfp_mask = orig_mask; in shrink_zones()
3026 vmpressure_prio(sc->gfp_mask, sc->target_mem_cgroup, in do_try_to_free_pages()
3149 static bool throttle_direct_reclaim(gfp_t gfp_mask, struct zonelist *zonelist, in throttle_direct_reclaim() argument
3188 gfp_zone(gfp_mask), nodemask) { in throttle_direct_reclaim()
3214 if (!(gfp_mask & __GFP_FS)) { in throttle_direct_reclaim()
3234 gfp_t gfp_mask, nodemask_t *nodemask) in try_to_free_pages() argument
3239 .gfp_mask = current_gfp_context(gfp_mask), in try_to_free_pages()
3240 .reclaim_idx = gfp_zone(gfp_mask), in try_to_free_pages()
3262 if (throttle_direct_reclaim(sc.gfp_mask, zonelist, nodemask)) in try_to_free_pages()
3266 trace_mm_vmscan_direct_reclaim_begin(order, sc.gfp_mask); in try_to_free_pages()
3280 gfp_t gfp_mask, bool noswap, in mem_cgroup_shrink_node() argument
3296 sc.gfp_mask = (gfp_mask & GFP_RECLAIM_MASK) | in mem_cgroup_shrink_node()
3300 sc.gfp_mask); in mem_cgroup_shrink_node()
3320 gfp_t gfp_mask, in try_to_free_mem_cgroup_pages() argument
3327 .gfp_mask = (current_gfp_context(gfp_mask) & GFP_RECLAIM_MASK) | in try_to_free_mem_cgroup_pages()
3341 struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask); in try_to_free_mem_cgroup_pages()
3344 trace_mm_vmscan_memcg_reclaim_begin(0, sc.gfp_mask); in try_to_free_mem_cgroup_pages()
3553 .gfp_mask = GFP_KERNEL, in balance_pgdat()
3664 sc.gfp_mask, &nr_soft_scanned); in balance_pgdat()
4005 .gfp_mask = GFP_HIGHUSER_MOVABLE, in shrink_all_memory()
4013 struct zonelist *zonelist = node_zonelist(numa_node_id(), sc.gfp_mask); in shrink_all_memory()
4017 fs_reclaim_acquire(sc.gfp_mask); in shrink_all_memory()
4025 fs_reclaim_release(sc.gfp_mask); in shrink_all_memory()
4156 static int __node_reclaim(struct pglist_data *pgdat, gfp_t gfp_mask, unsigned int order) in __node_reclaim() argument
4164 .gfp_mask = current_gfp_context(gfp_mask), in __node_reclaim()
4170 .reclaim_idx = gfp_zone(gfp_mask), in __node_reclaim()
4174 sc.gfp_mask); in __node_reclaim()
4177 fs_reclaim_acquire(sc.gfp_mask); in __node_reclaim()
4200 fs_reclaim_release(sc.gfp_mask); in __node_reclaim()
4207 int node_reclaim(struct pglist_data *pgdat, gfp_t gfp_mask, unsigned int order) in node_reclaim() argument
4229 if (!gfpflags_allow_blocking(gfp_mask) || (current->flags & PF_MEMALLOC)) in node_reclaim()
4244 ret = __node_reclaim(pgdat, gfp_mask, order); in node_reclaim()