Lines Matching refs:gfpflags
2188 static inline bool pfmemalloc_match(struct slab *slab, gfp_t gfpflags);
2955 slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid) in slab_out_of_memory() argument
2963 if ((gfpflags & __GFP_NOWARN) || !__ratelimit(&slub_oom_rs)) in slab_out_of_memory()
2967 nid, gfpflags, &gfpflags); in slab_out_of_memory()
2991 static inline bool pfmemalloc_match(struct slab *slab, gfp_t gfpflags) in pfmemalloc_match() argument
2994 return gfp_pfmemalloc_allowed(gfpflags); in pfmemalloc_match()
3052 static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, in ___slab_alloc() argument
3095 if (unlikely(!pfmemalloc_match(slab, gfpflags))) in ___slab_alloc()
3172 pc.flags = gfpflags; in ___slab_alloc()
3180 slab = new_slab(s, gfpflags, node); in ___slab_alloc()
3184 slab_out_of_memory(s, gfpflags, node); in ___slab_alloc()
3227 if (unlikely(!pfmemalloc_match(slab, gfpflags))) { in ___slab_alloc()
3265 static void *__slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, in __slab_alloc() argument
3279 p = ___slab_alloc(s, gfpflags, node, addr, c, orig_size); in __slab_alloc()
3309 gfp_t gfpflags, int node, unsigned long addr, size_t orig_size) in slab_alloc_node() argument
3318 s = slab_pre_alloc_hook(s, lru, &objcg, 1, gfpflags); in slab_alloc_node()
3322 object = kfence_alloc(s, orig_size, gfpflags); in slab_alloc_node()
3364 object = __slab_alloc(s, gfpflags, node, addr, c, orig_size); in slab_alloc_node()
3395 init = slab_want_init_on_alloc(gfpflags, s); in slab_alloc_node()
3398 slab_post_alloc_hook(s, objcg, gfpflags, 1, &object, init); in slab_alloc_node()
3404 gfp_t gfpflags, unsigned long addr, size_t orig_size) in slab_alloc() argument
3406 return slab_alloc_node(s, lru, gfpflags, NUMA_NO_NODE, addr, orig_size); in slab_alloc()
3411 gfp_t gfpflags) in __kmem_cache_alloc_lru() argument
3413 void *ret = slab_alloc(s, lru, gfpflags, _RET_IP_, s->object_size); in __kmem_cache_alloc_lru()
3415 trace_kmem_cache_alloc(_RET_IP_, ret, s, gfpflags, NUMA_NO_NODE); in __kmem_cache_alloc_lru()
3420 void *kmem_cache_alloc(struct kmem_cache *s, gfp_t gfpflags) in kmem_cache_alloc() argument
3422 return __kmem_cache_alloc_lru(s, NULL, gfpflags); in kmem_cache_alloc()
3427 gfp_t gfpflags) in kmem_cache_alloc_lru() argument
3429 return __kmem_cache_alloc_lru(s, lru, gfpflags); in kmem_cache_alloc_lru()
3433 void *__kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, in __kmem_cache_alloc_node() argument
3437 return slab_alloc_node(s, NULL, gfpflags, node, in __kmem_cache_alloc_node()
3441 void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, int node) in kmem_cache_alloc_node() argument
3443 void *ret = slab_alloc_node(s, NULL, gfpflags, node, _RET_IP_, s->object_size); in kmem_cache_alloc_node()
3445 trace_kmem_cache_alloc(_RET_IP_, ret, s, gfpflags, node); in kmem_cache_alloc_node()