Lines Matching refs:gfp_mask
412 struct page *__read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, in __read_swap_cache_async() argument
456 folio = vma_alloc_folio(gfp_mask, 0, vma, addr, false); in __read_swap_cache_async()
488 if (mem_cgroup_swapin_charge_folio(folio, NULL, gfp_mask, entry)) in __read_swap_cache_async()
492 if (add_to_swap_cache(folio, entry, gfp_mask & GFP_RECLAIM_MASK, &shadow)) in __read_swap_cache_async()
527 struct page *read_swap_cache_async(swp_entry_t entry, gfp_t gfp_mask, in read_swap_cache_async() argument
532 struct page *retpage = __read_swap_cache_async(entry, gfp_mask, in read_swap_cache_async()
620 struct page *swap_cluster_readahead(swp_entry_t entry, gfp_t gfp_mask, in swap_cluster_readahead() argument
652 gfp_mask, vma, addr, &page_allocated); in swap_cluster_readahead()
670 return read_swap_cache_async(entry, gfp_mask, vma, addr, NULL); in swap_cluster_readahead()
780 static struct page *swap_vma_readahead(swp_entry_t fentry, gfp_t gfp_mask, in swap_vma_readahead() argument
817 page = __read_swap_cache_async(entry, gfp_mask, vma, in swap_vma_readahead()
837 return read_swap_cache_async(fentry, gfp_mask, vma, vmf->address, in swap_vma_readahead()
853 struct page *swapin_readahead(swp_entry_t entry, gfp_t gfp_mask, in swapin_readahead() argument
857 swap_vma_readahead(entry, gfp_mask, vmf) : in swapin_readahead()
858 swap_cluster_readahead(entry, gfp_mask, vmf); in swapin_readahead()