Lines Matching refs:gfp_mask

213 static inline void warn_if_node_offline(int this_node, gfp_t gfp_mask)  in warn_if_node_offline()  argument
215 gfp_t warn_gfp = gfp_mask & (__GFP_THISNODE|__GFP_NOWARN); in warn_if_node_offline()
223 pr_warn("%pGg allocation from offline node %d\n", &gfp_mask, this_node); in warn_if_node_offline()
232 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument
235 warn_if_node_offline(nid, gfp_mask); in __alloc_pages_node()
237 return __alloc_pages(gfp_mask, order, nid, NULL); in __alloc_pages_node()
254 static inline struct page *alloc_pages_node(int nid, gfp_t gfp_mask, in alloc_pages_node() argument
260 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node()
269 static inline struct page *alloc_pages(gfp_t gfp_mask, unsigned int order) in alloc_pages() argument
271 return alloc_pages_node(numa_node_id(), gfp_mask, order); in alloc_pages()
280 #define alloc_page(gfp_mask) alloc_pages(gfp_mask, 0) argument
289 extern unsigned long __get_free_pages(gfp_t gfp_mask, unsigned int order);
290 extern unsigned long get_zeroed_page(gfp_t gfp_mask);
292 void *alloc_pages_exact(size_t size, gfp_t gfp_mask) __alloc_size(1);
294 __meminit void *alloc_pages_exact_nid(int nid, size_t size, gfp_t gfp_mask) __alloc_size(2);
296 #define __get_free_page(gfp_mask) \ argument
297 __get_free_pages((gfp_mask), 0)
299 #define __get_dma_pages(gfp_mask, order) \ argument
300 __get_free_pages((gfp_mask) | GFP_DMA, (order))
308 unsigned int fragsz, gfp_t gfp_mask,
312 unsigned int fragsz, gfp_t gfp_mask) in page_frag_alloc() argument
314 return page_frag_alloc_align(nc, fragsz, gfp_mask, ~0u); in page_frag_alloc()
339 bool gfp_pfmemalloc_allowed(gfp_t gfp_mask);
351 unsigned migratetype, gfp_t gfp_mask);
352 extern struct page *alloc_contig_pages(unsigned long nr_pages, gfp_t gfp_mask,