Home
last modified time | relevance | path

Searched refs:cache_size (Results 1 – 25 of 47) sorted by relevance

12

/Linux-v5.10/drivers/misc/lkdtm/
Dusercopy.c22 static volatile size_t cache_size = 1024; variable
220 memset(buf, 'B', cache_size); in do_usercopy_heap_whitelist()
223 offset = (cache_size / 4) + unconst; in do_usercopy_heap_whitelist()
224 size = (cache_size / 16) + unconst; in do_usercopy_heap_whitelist()
332 kmem_cache_create_usercopy("lkdtm-usercopy", cache_size, in lkdtm_usercopy_init()
334 cache_size / 4, in lkdtm_usercopy_init()
335 cache_size / 16, in lkdtm_usercopy_init()
/Linux-v5.10/drivers/infiniband/hw/hfi1/
Duser_pages.c55 static unsigned long cache_size = 256; variable
56 module_param(cache_size, ulong, S_IRUGO | S_IWUSR);
57 MODULE_PARM_DESC(cache_size, "Send and receive side cache size limit (in MB)");
75 size = (cache_size * (1UL << 20)); /* convert to bytes */ in hfi1_can_pin_pages()
/Linux-v5.10/tools/testing/selftests/resctrl/
Dcat_test.c23 unsigned long cache_size; variable
129 cache_size = 0; in cat_perf_miss_val()
146 ret = get_cache_size(cpu_no, cache_type, &cache_size); in cat_perf_miss_val()
149 printf("cache size :%lu\n", cache_size); in cat_perf_miss_val()
179 param.span = cache_size * (count_of_bits - n) / count_of_bits; in cat_perf_miss_val()
200 param.span = cache_size * n / count_of_bits; in cat_perf_miss_val()
Dcqm_test.c22 unsigned long cache_size; variable
118 cache_size = 0; in cqm_resctrl_val()
134 ret = get_cache_size(cpu_no, "L3", &cache_size); in cqm_resctrl_val()
137 printf("cache size :%lu\n", cache_size); in cqm_resctrl_val()
155 .span = cache_size * n / count_of_bits, in cqm_resctrl_val()
Dresctrlfs.c148 int get_cache_size(int cpu_no, char *cache_type, unsigned long *cache_size) in get_cache_size() argument
181 *cache_size = 0; in get_cache_size()
186 *cache_size = *cache_size * 10 + (cache_str[i] - '0'); in get_cache_size()
190 *cache_size = *cache_size * 1024; in get_cache_size()
194 *cache_size = *cache_size * 1024 * 1024; in get_cache_size()
Dfill_buf.c196 unsigned long long cache_size = span; in run_fill_buf() local
205 ret = fill_cache(cache_size, malloc_and_init_memory, memflush, op, in run_fill_buf()
Dresctrl.h96 int get_cache_size(int cpu_no, char *cache_type, unsigned long *cache_size);
/Linux-v5.10/drivers/mtd/
Dmtdblock.c30 unsigned int cache_size; member
88 mtdblk->cache_offset, mtdblk->cache_size); in write_cached_data()
91 mtdblk->cache_size, mtdblk->cache_data); in write_cached_data()
114 unsigned int sect_size = mtdblk->cache_size; in do_cached_write()
162 mtdblk->cache_size = sect_size; in do_cached_write()
184 unsigned int sect_size = mtdblk->cache_size; in do_cached_read()
237 if (unlikely(!mtdblk->cache_data && mtdblk->cache_size)) { in mtdblock_writesect()
265 mtdblk->cache_size = mbd->mtd->erasesize; in mtdblock_open()
/Linux-v5.10/arch/sh/kernel/cpu/
Dproc.c68 unsigned int cache_size; in show_cacheinfo() local
70 cache_size = info.ways * info.sets * info.linesz; in show_cacheinfo()
73 type, cache_size >> 10, info.ways); in show_cacheinfo()
/Linux-v5.10/arch/arm/mm/
Dcache-uniphier.c317 u32 level, cache_size; in __uniphier_cache_init() local
363 if (of_property_read_u32(np, "cache-size", &cache_size) || in __uniphier_cache_init()
364 cache_size == 0 || cache_size % (data->nsets * data->line_size)) { in __uniphier_cache_init()
371 data->way_mask = GENMASK(cache_size / data->nsets / data->line_size - 1, in __uniphier_cache_init()
Dcache-l2x0.c948 u32 cache_size = 0, sets = 0; in l2x0_cache_size_of_parse() local
954 of_property_read_u32(np, "cache-size", &cache_size); in l2x0_cache_size_of_parse()
959 if (!cache_size || !sets) in l2x0_cache_size_of_parse()
991 *associativity = cache_size / way_size; in l2x0_cache_size_of_parse()
999 cache_size, cache_size >> 10); in l2x0_cache_size_of_parse()
/Linux-v5.10/drivers/md/
Ddm-cache-policy-smq.c799 dm_cblock_t cache_size; member
1107 clear_bitset(mq->cache_hit_bits, from_cblock(mq->cache_size)); in end_cache_period()
1127 return from_cblock(mq->cache_size) * p / 100u; in percent_to_target()
1153 nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated; in free_target_met()
1719 static struct dm_cache_policy *__smq_create(dm_cblock_t cache_size, in __smq_create() argument
1734 mq->cache_size = cache_size; in __smq_create()
1737 calc_hotspot_params(origin_size, cache_block_size, from_cblock(cache_size), in __smq_create()
1742 if (space_init(&mq->es, total_sentinels + mq->nr_hotspot_blocks + from_cblock(cache_size))) { in __smq_create()
1760 total_sentinels + mq->nr_hotspot_blocks + from_cblock(cache_size)); in __smq_create()
1769 if (from_cblock(cache_size)) { in __smq_create()
[all …]
Ddm-cache-policy.c112 dm_cblock_t cache_size, in dm_cache_policy_create() argument
125 p = type->create(cache_size, origin_size, cache_block_size); in dm_cache_policy_create()
Ddm-cache-target.c402 dm_cblock_t cache_size; member
2368 cache->cache_size, in create_cache_policy()
2411 if (nr_blocks > (1 << 20) && cache->cache_size != size) in set_cache_size()
2417 cache->cache_size = size; in set_cache_size()
2471 dm_block_t cache_size = ca->cache_sectors; in cache_create() local
2474 cache_size = block_div(cache_size, ca->block_size); in cache_create()
2475 set_cache_size(cache, to_cblock(cache_size)); in cache_create()
2537 cache->dirty_bitset = alloc_bitset(from_cblock(cache->cache_size)); in cache_create()
2542 clear_bitset(cache->dirty_bitset, from_cblock(cache->cache_size)); in cache_create()
2736 r = dm_cache_set_dirty_bits(cache->cmd, from_cblock(cache->cache_size), cache->dirty_bitset); in write_dirty_bitset()
[all …]
Ddm-cache-policy.h176 struct dm_cache_policy *(*create)(dm_cblock_t cache_size,
Ddm-cache-policy-internal.h142 struct dm_cache_policy *dm_cache_policy_create(const char *name, dm_cblock_t cache_size,
/Linux-v5.10/drivers/gpu/drm/amd/amdkfd/
Dkfd_crat.c53 uint32_t cache_size; member
65 .cache_size = 16,
75 .cache_size = 16,
84 .cache_size = 8,
99 .cache_size = 16,
108 .cache_size = 8,
117 .cache_size = 4,
320 props->cache_size = cache->cache_size; in kfd_parse_subtype_cache()
578 pcache->cache_size = pcache_info[cache_type].cache_size; in fill_in_pcache()
Dkfd_topology.h127 uint32_t cache_size; member
/Linux-v5.10/arch/powerpc/kernel/
Dcacheinfo.c206 static int cache_size(const struct cache *cache, unsigned int *ret) in cache_size() function
209 const __be32 *cache_size; in cache_size() local
213 cache_size = of_get_property(cache->ofnode, propname, NULL); in cache_size()
214 if (!cache_size) in cache_size()
217 *ret = of_read_number(cache_size, 1); in cache_size()
225 if (cache_size(cache, &size)) in cache_size_kb()
290 if (cache_size(cache, &size)) in cache_associativity()
/Linux-v5.10/arch/microblaze/kernel/cpu/
Dcache.c92 #define CACHE_LOOP_LIMITS(start, end, cache_line_length, cache_size) \ argument
95 if (start < UINT_MAX - cache_size) \
96 end = min(start + cache_size, end); \
104 #define CACHE_ALL_LOOP(cache_size, line_length, op) \ argument
106 unsigned int len = cache_size - line_length; \
/Linux-v5.10/arch/mips/kernel/
Dpm-cps.c197 unsigned cache_size = cache->ways << cache->waybit; in cps_gen_cache_routine() local
209 if (cache_size < 0x8000) in cps_gen_cache_routine()
210 uasm_i_addiu(pp, t1, t0, cache_size); in cps_gen_cache_routine()
212 UASM_i_LA(pp, t1, (long)(CKSEG0 + cache_size)); in cps_gen_cache_routine()
/Linux-v5.10/net/openvswitch/
Dflow_table.c377 new->cache_size = size; in tbl_mask_cache_alloc()
378 if (new->cache_size > 0) { in tbl_mask_cache_alloc()
380 new->cache_size), in tbl_mask_cache_alloc()
396 if (size == mc->cache_size) in ovs_flow_tbl_masks_cache_resize()
801 if (unlikely(!skb_hash || mc->cache_size == 0)) { in ovs_flow_tbl_lookup_stats()
821 int index = hash & (mc->cache_size - 1); in ovs_flow_tbl_lookup_stats()
948 return READ_ONCE(mc->cache_size); in ovs_flow_tbl_masks_cache_size()
Dflow_table.h32 u32 cache_size; /* Must be ^2 value. */ member
/Linux-v5.10/drivers/block/
Dnull_blk.h55 unsigned long cache_size; /* disk cache size in MB */ member
Dnull_blk_main.c357 NULLB_DEVICE_ATTR(cache_size, ulong, NULL);
954 if ((nullb->dev->cache_size * 1024 * 1024) > in null_make_cache_space()
1105 nullb->dev->cache_size * 1024 * 1024); in null_handle_flush()
1743 dev->cache_size = 0; in null_validate_conf()
1744 dev->cache_size = min_t(unsigned long, ULONG_MAX / 1024 / 1024, in null_validate_conf()
1745 dev->cache_size); in null_validate_conf()
1847 if (dev->cache_size > 0) { in null_add_dev()

12