Lines Matching refs:oo
499 nr_slabs = DIV_ROUND_UP(nr_objects * 2, oo_objects(s->oo)); in slub_set_cpu_partial()
1863 struct kmem_cache_order_objects oo) in alloc_slab_page() argument
1867 unsigned int order = oo_order(oo); in alloc_slab_page()
1891 unsigned int count = oo_objects(s->oo); in init_cache_random_seq()
1961 freelist_count = oo_objects(s->oo); in shuffle_freelist()
1999 struct kmem_cache_order_objects oo = s->oo; in allocate_slab() local
2014 if ((alloc_gfp & __GFP_DIRECT_RECLAIM) && oo_order(oo) > oo_order(s->min)) in allocate_slab()
2017 slab = alloc_slab_page(alloc_gfp, node, oo); in allocate_slab()
2019 oo = s->min; in allocate_slab()
2025 slab = alloc_slab_page(alloc_gfp, node, oo); in allocate_slab()
2031 slab->objects = oo_objects(oo); in allocate_slab()
2035 account_slab(slab, oo_order(oo), s, flags); in allocate_slab()
2996 s->name, s->object_size, s->size, oo_order(s->oo), in slab_out_of_memory()
4503 s->oo = oo_make(order, size); in calculate_sizes()
4506 return !!oo_objects(s->oo); in calculate_sizes()
5210 obj_map = bitmap_alloc(oo_objects(s->oo), GFP_KERNEL); in validate_slab_cache()
5553 return sysfs_emit(buf, "%u\n", oo_objects(s->oo)); in objs_per_slab_show()
5559 return sysfs_emit(buf, "%u\n", oo_order(s->oo)); in order_show()
5662 objects = (slabs * oo_objects(s->oo)) / 2; in slabs_cpu_partial_show()
5672 objects = (slabs * oo_objects(s->oo)) / 2; in slabs_cpu_partial_show()
6400 obj_map = bitmap_alloc(oo_objects(s->oo), GFP_KERNEL); in slab_debug_trace_open()
6514 sinfo->objects_per_slab = oo_objects(s->oo); in get_slabinfo()
6515 sinfo->cache_order = oo_order(s->oo); in get_slabinfo()