Lines Matching refs:gfpflags
2267 static inline bool pfmemalloc_match(struct slab *slab, gfp_t gfpflags);
2983 slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid) in slab_out_of_memory() argument
2990 if ((gfpflags & __GFP_NOWARN) || !__ratelimit(&slub_oom_rs)) in slab_out_of_memory()
2994 nid, gfpflags, &gfpflags); in slab_out_of_memory()
3018 slab_out_of_memory(struct kmem_cache *s, gfp_t gfpflags, int nid) { } in slab_out_of_memory() argument
3021 static inline bool pfmemalloc_match(struct slab *slab, gfp_t gfpflags) in pfmemalloc_match() argument
3024 return gfp_pfmemalloc_allowed(gfpflags); in pfmemalloc_match()
3095 static void *___slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, in ___slab_alloc() argument
3138 if (unlikely(!pfmemalloc_match(slab, gfpflags))) in ___slab_alloc()
3215 pc.flags = gfpflags; in ___slab_alloc()
3223 slab = new_slab(s, gfpflags, node); in ___slab_alloc()
3227 slab_out_of_memory(s, gfpflags, node); in ___slab_alloc()
3270 if (unlikely(!pfmemalloc_match(slab, gfpflags))) { in ___slab_alloc()
3308 static void *__slab_alloc(struct kmem_cache *s, gfp_t gfpflags, int node, in __slab_alloc() argument
3322 p = ___slab_alloc(s, gfpflags, node, addr, c, orig_size); in __slab_alloc()
3330 gfp_t gfpflags, int node, unsigned long addr, size_t orig_size) in __slab_alloc_node() argument
3375 object = __slab_alloc(s, gfpflags, node, addr, c, orig_size); in __slab_alloc_node()
3405 gfp_t gfpflags, int node, unsigned long addr, size_t orig_size) in __slab_alloc_node() argument
3411 pc.flags = gfpflags; in __slab_alloc_node()
3419 slab = new_slab(s, gfpflags, node); in __slab_alloc_node()
3421 slab_out_of_memory(s, gfpflags, node); in __slab_alloc_node()
3454 gfp_t gfpflags, int node, unsigned long addr, size_t orig_size) in slab_alloc_node() argument
3460 s = slab_pre_alloc_hook(s, lru, &objcg, 1, gfpflags); in slab_alloc_node()
3464 object = kfence_alloc(s, orig_size, gfpflags); in slab_alloc_node()
3468 object = __slab_alloc_node(s, gfpflags, node, addr, orig_size); in slab_alloc_node()
3471 init = slab_want_init_on_alloc(gfpflags, s); in slab_alloc_node()
3478 slab_post_alloc_hook(s, objcg, gfpflags, 1, &object, init, orig_size); in slab_alloc_node()
3484 gfp_t gfpflags, unsigned long addr, size_t orig_size) in slab_alloc() argument
3486 return slab_alloc_node(s, lru, gfpflags, NUMA_NO_NODE, addr, orig_size); in slab_alloc()
3491 gfp_t gfpflags) in __kmem_cache_alloc_lru() argument
3493 void *ret = slab_alloc(s, lru, gfpflags, _RET_IP_, s->object_size); in __kmem_cache_alloc_lru()
3495 trace_kmem_cache_alloc(_RET_IP_, ret, s, gfpflags, NUMA_NO_NODE); in __kmem_cache_alloc_lru()
3500 void *kmem_cache_alloc(struct kmem_cache *s, gfp_t gfpflags) in kmem_cache_alloc() argument
3502 return __kmem_cache_alloc_lru(s, NULL, gfpflags); in kmem_cache_alloc()
3507 gfp_t gfpflags) in kmem_cache_alloc_lru() argument
3509 return __kmem_cache_alloc_lru(s, lru, gfpflags); in kmem_cache_alloc_lru()
3513 void *__kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, in __kmem_cache_alloc_node() argument
3517 return slab_alloc_node(s, NULL, gfpflags, node, in __kmem_cache_alloc_node()
3521 void *kmem_cache_alloc_node(struct kmem_cache *s, gfp_t gfpflags, int node) in kmem_cache_alloc_node() argument
3523 void *ret = slab_alloc_node(s, NULL, gfpflags, node, _RET_IP_, s->object_size); in kmem_cache_alloc_node()
3525 trace_kmem_cache_alloc(_RET_IP_, ret, s, gfpflags, node); in kmem_cache_alloc_node()