Lines Matching refs:gfp_mask

187 		      gfp_t gfp_mask, int node_id)  in mempool_init_node()  argument
197 gfp_mask, node_id); in mempool_init_node()
207 element = pool->alloc(gfp_mask, pool->pool_data); in mempool_init_node()
268 gfp_t gfp_mask, int node_id) in mempool_create_node() argument
272 pool = kzalloc_node(sizeof(*pool), gfp_mask, node_id); in mempool_create_node()
277 gfp_mask, node_id)) { in mempool_create_node()
380 void *mempool_alloc(mempool_t *pool, gfp_t gfp_mask) in mempool_alloc() argument
387 VM_WARN_ON_ONCE(gfp_mask & __GFP_ZERO); in mempool_alloc()
388 might_alloc(gfp_mask); in mempool_alloc()
390 gfp_mask |= __GFP_NOMEMALLOC; /* don't allocate emergency reserves */ in mempool_alloc()
391 gfp_mask |= __GFP_NORETRY; /* don't loop in __alloc_pages */ in mempool_alloc()
392 gfp_mask |= __GFP_NOWARN; /* failures are OK */ in mempool_alloc()
394 gfp_temp = gfp_mask & ~(__GFP_DIRECT_RECLAIM|__GFP_IO); in mempool_alloc()
420 if (gfp_temp != gfp_mask) { in mempool_alloc()
422 gfp_temp = gfp_mask; in mempool_alloc()
427 if (!(gfp_mask & __GFP_DIRECT_RECLAIM)) { in mempool_alloc()
514 void *mempool_alloc_slab(gfp_t gfp_mask, void *pool_data) in mempool_alloc_slab() argument
518 return kmem_cache_alloc(mem, gfp_mask); in mempool_alloc_slab()
533 void *mempool_kmalloc(gfp_t gfp_mask, void *pool_data) in mempool_kmalloc() argument
536 return kmalloc(size, gfp_mask); in mempool_kmalloc()
550 void *mempool_alloc_pages(gfp_t gfp_mask, void *pool_data) in mempool_alloc_pages() argument
553 return alloc_pages(gfp_mask, order); in mempool_alloc_pages()