| /Linux-v6.6/drivers/gpu/drm/i915/ |
| D | i915_active.c | 24 static struct kmem_cache *slab_cache; variable 172 kmem_cache_free(slab_cache, it); in __active_retire() 320 node = kmem_cache_alloc(slab_cache, GFP_ATOMIC); in active_instance() 766 kmem_cache_free(slab_cache, ref->cache); in i915_active_fini() 886 node = kmem_cache_alloc(slab_cache, GFP_KERNEL); in i915_active_acquire_preallocate_barrier() 934 kmem_cache_free(slab_cache, node); in i915_active_acquire_preallocate_barrier() 1198 kmem_cache_destroy(slab_cache); in i915_active_module_exit() 1203 slab_cache = KMEM_CACHE(active_node, SLAB_HWCACHE_ALIGN); in i915_active_module_init() 1204 if (!slab_cache) in i915_active_module_init()
|
| /Linux-v6.6/mm/ |
| D | slab.h | 47 struct kmem_cache *slab_cache; member 60 struct kmem_cache *slab_cache; member 107 SLAB_MATCH(compound_head, slab_cache); /* Ensure bit 0 is clear */ 629 return slab->slab_cache; in virt_to_cache()
|
| D | slab_common.c | 1074 s = slab->slab_cache; in kfree() 1109 skip_orig_size_check(folio_slab(folio)->slab_cache, object); in __ksize() 1112 return slab_ksize(folio_slab(folio)->slab_cache); in __ksize()
|
| D | slub.c | 1486 if (unlikely(s != slab->slab_cache)) { in free_consistency_checks() 1490 } else if (!slab->slab_cache) { in free_consistency_checks() 2037 slab->slab_cache = s; in allocate_slab() 2094 __free_slab(slab->slab_cache, slab); in rcu_free_slab() 3875 df->s = df->slab->slab_cache; in build_detached_freelist() 4655 struct kmem_cache *s = slab->slab_cache; in __kmem_obj_info() 4757 s = slab->slab_cache; in __check_heap_object() 5011 p->slab_cache = s; in bootstrap() 5015 p->slab_cache = s; in bootstrap()
|
| D | slab.c | 1402 cachep = slab->slab_cache; in kmem_rcu_free() 2581 slab->slab_cache = cachep; in cache_grow_begin() 3534 cachep = slab->slab_cache; in __kmem_obj_info() 3605 s = folio_slab(folio)->slab_cache; in kmem_cache_free_bulk() 4008 cachep = slab->slab_cache; in __check_heap_object()
|
| D | memcontrol.c | 2940 off = obj_to_index(slab->slab_cache, slab, p); in mem_cgroup_from_obj_folio()
|
| /Linux-v6.6/Documentation/translations/zh_CN/mm/ |
| D | split_page_table_lock.rst | 62 确保架构不使用slab分配器来分配页表:slab使用page->slab_cache来分配其页
|
| /Linux-v6.6/mm/kasan/ |
| D | common.c | 294 ____kasan_slab_free(slab->slab_cache, ptr, ip, false, false); in __kasan_slab_free_mempool() 442 return ____kasan_kmalloc(slab->slab_cache, object, size, flags); in __kasan_krealloc()
|
| D | quarantine.c | 129 return virt_to_slab(qlink)->slab_cache; in qlink_to_cache()
|
| D | report.c | 495 info->cache = slab->slab_cache; in complete_report_info()
|
| D | generic.c | 485 cache = slab->slab_cache; in __kasan_record_aux_stack()
|
| /Linux-v6.6/drivers/md/ |
| D | dm-bufio.c | 956 struct kmem_cache *slab_cache; member 1146 if (unlikely(c->slab_cache != NULL)) { in alloc_buffer_data() 1148 return kmem_cache_alloc(c->slab_cache, gfp_mask); in alloc_buffer_data() 1171 kmem_cache_free(c->slab_cache, data); in free_buffer_data() 2462 c->slab_cache = kmem_cache_create(slab_name, block_size, align, in dm_bufio_client_create() 2464 if (!c->slab_cache) { in dm_bufio_client_create() 2517 kmem_cache_destroy(c->slab_cache); in dm_bufio_client_create() 2568 kmem_cache_destroy(c->slab_cache); in dm_bufio_client_destroy()
|
| D | raid5.h | 637 struct kmem_cache *slab_cache; /* for allocating stripes */ member
|
| D | raid5.c | 2408 sh = alloc_stripe(conf->slab_cache, gfp, conf->pool_size, conf); in grow_one_stripe() 2414 free_stripe(conf->slab_cache, sh); in grow_one_stripe() 2447 conf->slab_cache = sc; in grow_stripes() 2622 free_stripe(conf->slab_cache, osh); in resize_stripes() 2630 kmem_cache_destroy(conf->slab_cache); in resize_stripes() 2660 conf->slab_cache = sc; in resize_stripes() 2718 free_stripe(conf->slab_cache, sh); in drop_one_stripe() 2730 kmem_cache_destroy(conf->slab_cache); in shrink_stripes() 2731 conf->slab_cache = NULL; in shrink_stripes()
|
| /Linux-v6.6/tools/cgroup/ |
| D | memcg_slabinfo.py | 199 cache = slab.slab_cache
|
| /Linux-v6.6/Documentation/mm/ |
| D | split_page_table_lock.rst | 65 allocation: slab uses page->slab_cache for its pages.
|
| /Linux-v6.6/mm/kfence/ |
| D | core.c | 465 slab->slab_cache = cache; in kfence_guarded_alloc()
|