/Linux-v5.10/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_icm_pool.c | 43 struct mlx5dr_icm_bucket *buckets; member 335 hot_size += pool->buckets[chunk_order].hot_list_count * in dr_icm_hot_mem_size() 376 bool buckets[DR_CHUNK_SIZE_MAX]) in dr_icm_chill_buckets_start() 382 bucket = &pool->buckets[i]; in dr_icm_chill_buckets_start() 393 buckets[i] = true; in dr_icm_chill_buckets_start() 400 bool buckets[DR_CHUNK_SIZE_MAX]) in dr_icm_chill_buckets_end() 406 bucket = &pool->buckets[i]; in dr_icm_chill_buckets_end() 412 if (!buckets[i]) in dr_icm_chill_buckets_end() 422 bool buckets[DR_CHUNK_SIZE_MAX]) in dr_icm_chill_buckets_abort() 428 bucket = &pool->buckets[i]; in dr_icm_chill_buckets_abort() [all …]
|
/Linux-v5.10/drivers/s390/scsi/ |
D | zfcp_reqlist.h | 24 struct list_head buckets[ZFCP_REQ_LIST_BUCKETS]; member 50 INIT_LIST_HEAD(&rl->buckets[i]); in zfcp_reqlist_alloc() 66 if (!list_empty(&rl->buckets[i])) in zfcp_reqlist_isempty() 90 list_for_each_entry(req, &rl->buckets[i], list) in _zfcp_reqlist_find() 163 list_add_tail(&req->list, &rl->buckets[i]); in zfcp_reqlist_add() 180 list_splice_init(&rl->buckets[i], list); in zfcp_reqlist_move() 207 list_for_each_entry(req, &rl->buckets[i], list) in zfcp_reqlist_apply_for_all()
|
/Linux-v5.10/tools/lib/bpf/ |
D | hashmap.h | 57 struct hashmap_entry **buckets; member 67 .buckets = NULL, \ 157 for (cur = map->buckets[bkt]; cur; cur = cur->next) 169 for (cur = map->buckets[bkt]; \ 180 for (cur = map->buckets \ 181 ? map->buckets[hash_bits(map->hash_fn((_key), map->ctx), map->cap_bits)] \ 188 for (cur = map->buckets \ 189 ? map->buckets[hash_bits(map->hash_fn((_key), map->ctx), map->cap_bits)] \
|
D | hashmap.c | 45 map->buckets = NULL; in hashmap__init() 71 free(map->buckets); in hashmap__clear() 72 map->buckets = NULL; in hashmap__clear() 124 free(map->buckets); in hashmap_grow() 125 map->buckets = new_buckets; in hashmap_grow() 137 if (!map->buckets) in hashmap_find_entry() 140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry() 200 hashmap_add_entry(&map->buckets[h], entry); in hashmap__insert()
|
/Linux-v5.10/tools/perf/util/ |
D | hashmap.h | 57 struct hashmap_entry **buckets; member 67 .buckets = NULL, \ 157 for (cur = map->buckets[bkt]; cur; cur = cur->next) 169 for (cur = map->buckets[bkt]; \ 180 for (cur = map->buckets \ 181 ? map->buckets[hash_bits(map->hash_fn((_key), map->ctx), map->cap_bits)] \ 188 for (cur = map->buckets \ 189 ? map->buckets[hash_bits(map->hash_fn((_key), map->ctx), map->cap_bits)] \
|
D | hashmap.c | 45 map->buckets = NULL; in hashmap__init() 71 free(map->buckets); in hashmap__clear() 72 map->buckets = NULL; in hashmap__clear() 124 free(map->buckets); in hashmap_grow() 125 map->buckets = new_buckets; in hashmap_grow() 137 if (!map->buckets) in hashmap_find_entry() 140 for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; in hashmap_find_entry() 200 hashmap_add_entry(&map->buckets[h], entry); in hashmap__insert()
|
/Linux-v5.10/block/ |
D | blk-stat.c | 86 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn() 93 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn() 105 unsigned int buckets, void *data) in blk_stat_alloc_callback() argument 113 cb->stat = kmalloc_array(buckets, sizeof(struct blk_rq_stat), in blk_stat_alloc_callback() 119 cb->cpu_stat = __alloc_percpu(buckets * sizeof(struct blk_rq_stat), in blk_stat_alloc_callback() 130 cb->buckets = buckets; in blk_stat_alloc_callback() 147 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_add_callback()
|
D | blk-stat.h | 45 unsigned int buckets; member 87 unsigned int buckets, void *data);
|
D | kyber-iosched.c | 134 atomic_t buckets[KYBER_OTHER][2][KYBER_LATENCY_BUCKETS]; member 213 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in flush_latency_buckets() local 214 atomic_t *cpu_buckets = cpu_latency->buckets[sched_domain][type]; in flush_latency_buckets() 218 buckets[bucket] += atomic_xchg(&cpu_buckets[bucket], 0); in flush_latency_buckets() 229 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in calculate_percentile() local 233 samples += buckets[bucket]; in calculate_percentile() 252 if (buckets[bucket] >= percentile_samples) in calculate_percentile() 254 percentile_samples -= buckets[bucket]; in calculate_percentile() 256 memset(buckets, 0, sizeof(kqd->latency_buckets[sched_domain][type])); in calculate_percentile() 632 atomic_inc(&cpu_latency->buckets[sched_domain][type][bucket]); in add_latency_sample()
|
/Linux-v5.10/net/ceph/crush/ |
D | crush.c | 111 if (map->buckets) { in crush_destroy() 114 if (map->buckets[b] == NULL) in crush_destroy() 116 crush_destroy_bucket(map->buckets[b]); in crush_destroy() 118 kfree(map->buckets); in crush_destroy()
|
D | mapper.c | 527 itemtype = map->buckets[-1-item]->type; in crush_choose_firstn() 540 in = map->buckets[-1-item]; in crush_choose_firstn() 564 map->buckets[-1-item], in crush_choose_firstn() 741 itemtype = map->buckets[-1-item]->type; in crush_choose_indep() 758 in = map->buckets[-1-item]; in crush_choose_indep() 778 map->buckets[-1-item], in crush_choose_indep() 865 if (!map->buckets[b]) in crush_init_workspace() 869 switch (map->buckets[b]->alg) { in crush_init_workspace() 877 v += map->buckets[b]->size * sizeof(__u32); in crush_init_workspace() 949 map->buckets[-1-curstep->arg1])) { in crush_do_rule() [all …]
|
/Linux-v5.10/net/netfilter/ipvs/ |
D | ip_vs_sh.c | 70 struct ip_vs_sh_bucket buckets[IP_VS_SH_TAB_SIZE]; member 108 struct ip_vs_dest *dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get() 130 dest = rcu_dereference(s->buckets[ihash].dest); in ip_vs_sh_get_fallback() 145 dest = rcu_dereference(s->buckets[hash].dest); in ip_vs_sh_get_fallback() 172 b = &s->buckets[0]; in ip_vs_sh_reassign() 216 b = &s->buckets[0]; in ip_vs_sh_flush()
|
D | ip_vs_dh.c | 64 struct ip_vs_dh_bucket buckets[IP_VS_DH_TAB_SIZE]; member 90 return rcu_dereference(s->buckets[ip_vs_dh_hashkey(af, addr)].dest); in ip_vs_dh_get() 106 b = &s->buckets[0]; in ip_vs_dh_reassign() 140 b = &s->buckets[0]; in ip_vs_dh_flush()
|
/Linux-v5.10/kernel/bpf/ |
D | bpf_local_storage.c | 21 return &smap->buckets[hash_ptr(selem, smap->bucket_log)]; in select_bucket() 491 b = &smap->buckets[i]; in bpf_local_storage_map_free() 518 kvfree(smap->buckets); in bpf_local_storage_map_free() 558 cost = sizeof(*smap->buckets) * nbuckets + sizeof(*smap); in bpf_local_storage_map_alloc() 566 smap->buckets = kvcalloc(sizeof(*smap->buckets), nbuckets, in bpf_local_storage_map_alloc() 568 if (!smap->buckets) { in bpf_local_storage_map_alloc() 575 INIT_HLIST_HEAD(&smap->buckets[i].list); in bpf_local_storage_map_alloc() 576 raw_spin_lock_init(&smap->buckets[i].lock); in bpf_local_storage_map_alloc()
|
D | stackmap.c | 32 struct stack_map_bucket *buckets[]; member 423 bucket = READ_ONCE(smap->buckets[id]); in __bpf_get_stackid() 467 old_bucket = xchg(&smap->buckets[id], new_bucket); in __bpf_get_stackid() 760 bucket = xchg(&smap->buckets[id], NULL); in bpf_stackmap_copy() 768 old_bucket = xchg(&smap->buckets[id], bucket); in bpf_stackmap_copy() 787 if (id >= smap->n_buckets || !smap->buckets[id]) in stack_map_get_next_key() 793 while (id < smap->n_buckets && !smap->buckets[id]) in stack_map_get_next_key() 819 old_bucket = xchg(&smap->buckets[id], NULL); in stack_map_delete_elem()
|
/Linux-v5.10/fs/nfs/ |
D | pnfs_nfs.c | 106 p = kmalloc(struct_size(p, buckets, n), gfp_flags); in pnfs_alloc_commit_array() 113 for (b = &p->buckets[0]; n != 0; b++, n--) { in pnfs_alloc_commit_array() 263 struct pnfs_commit_bucket *buckets, in pnfs_bucket_scan_array() argument 271 cnt = pnfs_bucket_scan_ds_commit_list(&buckets[i], cinfo, max); in pnfs_bucket_scan_array() 292 cnt = pnfs_bucket_scan_array(cinfo, array->buckets, in pnfs_generic_scan_commit_lists() 308 struct pnfs_commit_bucket *buckets, in pnfs_bucket_recover_commit_reqs() argument 318 for (i = 0, b = buckets; i < nbuckets; i++, b++) { in pnfs_bucket_recover_commit_reqs() 347 array->buckets, in pnfs_generic_recover_commit_reqs() 359 pnfs_bucket_search_commit_reqs(struct pnfs_commit_bucket *buckets, in pnfs_bucket_search_commit_reqs() argument 368 for (i = 0, b = buckets; i < nbuckets; i++, b++) { in pnfs_bucket_search_commit_reqs() [all …]
|
D | nfs42xattr.c | 70 struct nfs4_xattr_bucket buckets[NFS4_XATTR_HASH_SIZE]; member 111 INIT_HLIST_HEAD(&cache->buckets[i].hlist); in nfs4_xattr_hash_init() 112 spin_lock_init(&cache->buckets[i].lock); in nfs4_xattr_hash_init() 113 cache->buckets[i].cache = cache; in nfs4_xattr_hash_init() 114 cache->buckets[i].draining = false; in nfs4_xattr_hash_init() 276 if (WARN_ON(!hlist_empty(&cache->buckets[i].hlist))) in nfs4_xattr_free_cache_cb() 278 cache->buckets[i].draining = false; in nfs4_xattr_free_cache_cb() 395 bucket = &cache->buckets[i]; in nfs4_xattr_discard_cache() 507 return &cache->buckets[jhash(name, strlen(name), 0) & in nfs4_xattr_hash_bucket() 508 (ARRAY_SIZE(cache->buckets) - 1)]; in nfs4_xattr_hash_bucket() [all …]
|
/Linux-v5.10/drivers/net/ethernet/mellanox/mlx5/core/ipoib/ |
D | ipoib_vlan.c | 45 struct hlist_head buckets[1 << MLX5I_MAX_LOG_PKEY_SUP]; member 71 static struct qpn_to_netdev *mlx5i_find_qpn_to_netdev_node(struct hlist_head *buckets, in mlx5i_find_qpn_to_netdev_node() argument 74 struct hlist_head *h = &buckets[hash_32(qpn, MLX5I_MAX_LOG_PKEY_SUP)]; in mlx5i_find_qpn_to_netdev_node() 99 hlist_add_head(&new_node->hlist, &ht->buckets[key]); in mlx5i_pkey_add_qpn() 112 node = mlx5i_find_qpn_to_netdev_node(ht->buckets, qpn); in mlx5i_pkey_del_qpn() 131 node = mlx5i_find_qpn_to_netdev_node(ipriv->qpn_htbl->buckets, qpn); in mlx5i_pkey_get_netdev()
|
/Linux-v5.10/drivers/md/ |
D | dm-region-hash.c | 70 struct list_head *buckets; member 206 rh->buckets = vmalloc(array_size(nr_buckets, sizeof(*rh->buckets))); in dm_region_hash_create() 207 if (!rh->buckets) { in dm_region_hash_create() 214 INIT_LIST_HEAD(rh->buckets + i); in dm_region_hash_create() 228 vfree(rh->buckets); in dm_region_hash_create() 244 list_for_each_entry_safe(reg, nreg, rh->buckets + h, in dm_region_hash_destroy() 255 vfree(rh->buckets); in dm_region_hash_destroy() 274 struct list_head *bucket = rh->buckets + rh_hash(rh, region); in __rh_lookup() 285 list_add(®->hash_list, rh->buckets + rh_hash(rh, reg->key)); in __rh_insert()
|
/Linux-v5.10/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
D | pno.c | 298 struct brcmf_gscan_bucket_config **buckets, in brcmf_pno_prep_fwconfig() argument 323 *buckets = NULL; in brcmf_pno_prep_fwconfig() 355 *buckets = fw_buckets; in brcmf_pno_prep_fwconfig() 396 struct brcmf_gscan_bucket_config *buckets; in brcmf_pno_config_sched_scans() local 403 n_buckets = brcmf_pno_prep_fwconfig(pi, &pno_cfg, &buckets, in brcmf_pno_config_sched_scans() 408 gsz = sizeof(*gscan_cfg) + (n_buckets - 1) * sizeof(*buckets); in brcmf_pno_config_sched_scans() 437 memcpy(&gscan_cfg->bucket[0], buckets, in brcmf_pno_config_sched_scans() 438 n_buckets * sizeof(*buckets)); in brcmf_pno_config_sched_scans() 463 kfree(buckets); in brcmf_pno_config_sched_scans()
|
/Linux-v5.10/drivers/md/persistent-data/ |
D | dm-transaction-manager.c | 96 struct hlist_head buckets[DM_HASH_SIZE]; member 110 hlist_for_each_entry(si, tm->buckets + bucket, hlist) in is_shadow() 134 hlist_add_head(&si->hlist, tm->buckets + bucket); in insert_shadow() 148 bucket = tm->buckets + i; in wipe_shadow_table() 177 INIT_HLIST_HEAD(tm->buckets + i); in dm_tm_create()
|
/Linux-v5.10/net/sched/ |
D | sch_hhf.c | 128 struct wdrr_bucket buckets[WDRR_BUCKET_CNT]; member 355 bucket = &q->buckets[WDRR_BUCKET_FOR_HH]; in hhf_drop() 357 bucket = &q->buckets[WDRR_BUCKET_FOR_NON_HH]; in hhf_drop() 368 return bucket - q->buckets; in hhf_drop() 381 bucket = &q->buckets[idx]; in hhf_enqueue() 435 int weight = (bucket - q->buckets == WDRR_BUCKET_FOR_HH) ? in hhf_dequeue() 645 struct wdrr_bucket *bucket = q->buckets + i; in hhf_init()
|
/Linux-v5.10/net/netfilter/ |
D | nft_set_hash.c | 413 u32 buckets; member 431 hash = reciprocal_scale(hash, priv->buckets); in nft_hash_lookup() 451 hash = reciprocal_scale(hash, priv->buckets); in nft_hash_get() 471 hash = reciprocal_scale(hash, priv->buckets); in nft_hash_lookup_fast() 495 hash = reciprocal_scale(hash, priv->buckets); in nft_jhash() 577 for (i = 0; i < priv->buckets; i++) { in nft_hash_walk() 608 priv->buckets = nft_hash_buckets(desc->size); in nft_hash_init() 621 for (i = 0; i < priv->buckets; i++) { in nft_hash_destroy()
|
/Linux-v5.10/drivers/md/bcache/ |
D | alloc.c | 146 trace_bcache_invalidate(ca, b - ca->buckets); in __bch_invalidate_one_bucket() 157 fifo_push(&ca->free_inc, b - ca->buckets); in bch_invalidate_one_bucket() 226 b = ca->buckets + ca->fifo_last_bucket++; in invalidate_buckets_fifo() 252 b = ca->buckets + n; in invalidate_buckets_random() 445 b = ca->buckets + r; in bch_bucket_alloc() 508 k->ptr[0] = MAKE_PTR(ca->buckets[b].gen, in __bch_bucket_alloc_set()
|
/Linux-v5.10/Documentation/userspace-api/media/v4l/ |
D | pixfmt-meta-vsp1-hgt.rst | 23 The histogram is a matrix of 6 Hue and 32 Saturation buckets, 192 in 24 total. Each HSV value is added to one or more buckets with a weight 26 corresponding buckets is done by inspecting the H and S value independently.
|