/Linux-v5.10/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_icm_pool.c | 181 struct mlx5dr_icm_bucket *bucket = chunk->bucket; in dr_icm_chunk_ste_init() local 183 chunk->ste_arr = kvzalloc(bucket->num_of_entries * in dr_icm_chunk_ste_init() 188 chunk->hw_ste_arr = kvzalloc(bucket->num_of_entries * in dr_icm_chunk_ste_init() 193 chunk->miss_list = kvmalloc(bucket->num_of_entries * in dr_icm_chunk_ste_init() 207 static int dr_icm_chunks_create(struct mlx5dr_icm_bucket *bucket) in dr_icm_chunks_create() argument 210 struct mlx5dr_icm_pool *pool = bucket->pool; in dr_icm_chunks_create() 215 mr_req_size = bucket->num_of_entries * bucket->entry_size; in dr_icm_chunks_create() 243 chunk->bucket = bucket; in dr_icm_chunks_create() 249 chunk->num_of_entries = bucket->num_of_entries; in dr_icm_chunks_create() 250 chunk->byte_size = chunk->num_of_entries * bucket->entry_size; in dr_icm_chunks_create() [all …]
|
/Linux-v5.10/net/mptcp/ |
D | token.c | 112 struct token_bucket *bucket; in mptcp_token_new_request() local 123 bucket = token_bucket(token); in mptcp_token_new_request() 124 spin_lock_bh(&bucket->lock); in mptcp_token_new_request() 125 if (__token_bucket_busy(bucket, token)) { in mptcp_token_new_request() 126 spin_unlock_bh(&bucket->lock); in mptcp_token_new_request() 130 hlist_nulls_add_head_rcu(&subflow_req->token_node, &bucket->req_chain); in mptcp_token_new_request() 131 bucket->chain_len++; in mptcp_token_new_request() 132 spin_unlock_bh(&bucket->lock); in mptcp_token_new_request() 157 struct token_bucket *bucket; in mptcp_token_new_connect() local 166 bucket = token_bucket(subflow->token); in mptcp_token_new_connect() [all …]
|
/Linux-v5.10/net/ceph/crush/ |
D | mapper.c | 74 static int bucket_perm_choose(const struct crush_bucket *bucket, in bucket_perm_choose() argument 78 unsigned int pr = r % bucket->size; in bucket_perm_choose() 83 dprintk("bucket %d new x=%d\n", bucket->id, x); in bucket_perm_choose() 88 s = crush_hash32_3(bucket->hash, x, bucket->id, 0) % in bucket_perm_choose() 89 bucket->size; in bucket_perm_choose() 95 for (i = 0; i < bucket->size; i++) in bucket_perm_choose() 100 for (i = 1; i < bucket->size; i++) in bucket_perm_choose() 112 if (p < bucket->size - 1) { in bucket_perm_choose() 113 i = crush_hash32_3(bucket->hash, x, bucket->id, p) % in bucket_perm_choose() 114 (bucket->size - p); in bucket_perm_choose() [all …]
|
/Linux-v5.10/block/ |
D | blk-stat.c | 56 int bucket, cpu; in blk_stat_add() local 69 bucket = cb->bucket_fn(rq); in blk_stat_add() 70 if (bucket < 0) in blk_stat_add() 73 stat = &per_cpu_ptr(cb->cpu_stat, cpu)[bucket]; in blk_stat_add() 83 unsigned int bucket; in blk_stat_timer_fn() local 86 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn() 87 blk_rq_stat_init(&cb->stat[bucket]); in blk_stat_timer_fn() 93 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn() 94 blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); in blk_stat_timer_fn() 95 blk_rq_stat_init(&cpu_stat[bucket]); in blk_stat_timer_fn() [all …]
|
D | kyber-iosched.c | 215 unsigned int bucket; in flush_latency_buckets() local 217 for (bucket = 0; bucket < KYBER_LATENCY_BUCKETS; bucket++) in flush_latency_buckets() 218 buckets[bucket] += atomic_xchg(&cpu_buckets[bucket], 0); in flush_latency_buckets() 230 unsigned int bucket, samples = 0, percentile_samples; in calculate_percentile() local 232 for (bucket = 0; bucket < KYBER_LATENCY_BUCKETS; bucket++) in calculate_percentile() 233 samples += buckets[bucket]; in calculate_percentile() 251 for (bucket = 0; bucket < KYBER_LATENCY_BUCKETS - 1; bucket++) { in calculate_percentile() 252 if (buckets[bucket] >= percentile_samples) in calculate_percentile() 254 percentile_samples -= buckets[bucket]; in calculate_percentile() 260 bucket + 1, 1 << KYBER_LATENCY_SHIFT, samples); in calculate_percentile() [all …]
|
/Linux-v5.10/net/sched/ |
D | sch_hhf.c | 329 static struct sk_buff *dequeue_head(struct wdrr_bucket *bucket) in dequeue_head() argument 331 struct sk_buff *skb = bucket->head; in dequeue_head() 333 bucket->head = skb->next; in dequeue_head() 339 static void bucket_add(struct wdrr_bucket *bucket, struct sk_buff *skb) in bucket_add() argument 341 if (bucket->head == NULL) in bucket_add() 342 bucket->head = skb; in bucket_add() 344 bucket->tail->next = skb; in bucket_add() 345 bucket->tail = skb; in bucket_add() 352 struct wdrr_bucket *bucket; in hhf_drop() local 355 bucket = &q->buckets[WDRR_BUCKET_FOR_HH]; in hhf_drop() [all …]
|
/Linux-v5.10/drivers/infiniband/sw/rdmavt/ |
D | trace_qp.h | 60 TP_PROTO(struct rvt_qp *qp, u32 bucket), 61 TP_ARGS(qp, bucket), 65 __field(u32, bucket) 70 __entry->bucket = bucket; 76 __entry->bucket 81 TP_PROTO(struct rvt_qp *qp, u32 bucket), 82 TP_ARGS(qp, bucket)); 85 TP_PROTO(struct rvt_qp *qp, u32 bucket), 86 TP_ARGS(qp, bucket));
|
/Linux-v5.10/drivers/interconnect/qcom/ |
D | bcm-voter.c | 71 size_t i, bucket; in bcm_aggregate() local 76 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate() 79 temp = bcm_div(node->sum_avg[bucket] * bcm->aux_data.width, in bcm_aggregate() 81 agg_avg[bucket] = max(agg_avg[bucket], temp); in bcm_aggregate() 83 temp = bcm_div(node->max_peak[bucket] * bcm->aux_data.width, in bcm_aggregate() 85 agg_peak[bucket] = max(agg_peak[bucket], temp); in bcm_aggregate() 88 temp = agg_avg[bucket] * bcm->vote_scale; in bcm_aggregate() 89 bcm->vote_x[bucket] = bcm_div(temp, bcm->aux_data.unit); in bcm_aggregate() 91 temp = agg_peak[bucket] * bcm->vote_scale; in bcm_aggregate() 92 bcm->vote_y[bucket] = bcm_div(temp, bcm->aux_data.unit); in bcm_aggregate() [all …]
|
/Linux-v5.10/fs/nfs/ |
D | nfs42xattr.c | 87 struct nfs4_xattr_bucket *bucket; member 238 entry->bucket = NULL; in nfs4_xattr_alloc_entry() 389 struct nfs4_xattr_bucket *bucket; in nfs4_xattr_discard_cache() local 395 bucket = &cache->buckets[i]; in nfs4_xattr_discard_cache() 397 spin_lock(&bucket->lock); in nfs4_xattr_discard_cache() 398 bucket->draining = true; in nfs4_xattr_discard_cache() 399 hlist_for_each_entry_safe(entry, n, &bucket->hlist, hnode) { in nfs4_xattr_discard_cache() 404 spin_unlock(&bucket->lock); in nfs4_xattr_discard_cache() 512 nfs4_xattr_get_entry(struct nfs4_xattr_bucket *bucket, const char *name) in nfs4_xattr_get_entry() argument 518 hlist_for_each_entry(entry, &bucket->hlist, hnode) { in nfs4_xattr_get_entry() [all …]
|
D | pnfs_nfs.c | 63 pnfs_free_bucket_lseg(struct pnfs_commit_bucket *bucket) in pnfs_free_bucket_lseg() argument 65 if (list_empty(&bucket->committing) && list_empty(&bucket->written)) { in pnfs_free_bucket_lseg() 66 struct pnfs_layout_segment *freeme = bucket->lseg; in pnfs_free_bucket_lseg() 67 bucket->lseg = NULL; in pnfs_free_bucket_lseg() 87 struct pnfs_commit_bucket *bucket; in pnfs_generic_clear_request_commit() local 89 bucket = list_first_entry(&req->wb_list, in pnfs_generic_clear_request_commit() 92 freeme = pnfs_free_bucket_lseg(bucket); in pnfs_generic_clear_request_commit() 245 pnfs_bucket_scan_ds_commit_list(struct pnfs_commit_bucket *bucket, in pnfs_bucket_scan_ds_commit_list() argument 249 struct list_head *src = &bucket->written; in pnfs_bucket_scan_ds_commit_list() 250 struct list_head *dst = &bucket->committing; in pnfs_bucket_scan_ds_commit_list() [all …]
|
/Linux-v5.10/net/9p/ |
D | error.c | 181 int bucket; in p9_error_init() local 184 for (bucket = 0; bucket < ERRHASHSZ; bucket++) in p9_error_init() 185 INIT_HLIST_HEAD(&hash_errmap[bucket]); in p9_error_init() 190 bucket = jhash(c->name, c->namelen, 0) % ERRHASHSZ; in p9_error_init() 192 hlist_add_head(&c->list, &hash_errmap[bucket]); in p9_error_init() 210 int bucket; in p9_errstr2errno() local 214 bucket = jhash(errstr, len, 0) % ERRHASHSZ; in p9_errstr2errno() 215 hlist_for_each_entry(c, &hash_errmap[bucket], list) { in p9_errstr2errno()
|
/Linux-v5.10/fs/dlm/ |
D | debug_fs.c | 368 unsigned bucket; member 427 unsigned bucket, entry; in table_seq_start() local 430 bucket = n >> 32; in table_seq_start() 433 if (bucket >= ls->ls_rsbtbl_size) in table_seq_start() 450 tree = toss ? &ls->ls_rsbtbl[bucket].toss : &ls->ls_rsbtbl[bucket].keep; in table_seq_start() 452 spin_lock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start() 459 ri->bucket = bucket; in table_seq_start() 460 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start() 465 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start() 475 bucket++; in table_seq_start() [all …]
|
/Linux-v5.10/net/vmw_vsock/ |
D | diag.c | 52 unsigned int bucket; in vsock_diag_dump() local 63 bucket = cb->args[1]; in vsock_diag_dump() 72 while (bucket < ARRAY_SIZE(vsock_bind_table)) { in vsock_diag_dump() 73 struct list_head *head = &vsock_bind_table[bucket]; in vsock_diag_dump() 94 bucket++; in vsock_diag_dump() 98 bucket = 0; in vsock_diag_dump() 102 while (bucket < ARRAY_SIZE(vsock_connected_table)) { in vsock_diag_dump() 103 struct list_head *head = &vsock_connected_table[bucket]; in vsock_diag_dump() 128 bucket++; in vsock_diag_dump() 135 cb->args[1] = bucket; in vsock_diag_dump()
|
/Linux-v5.10/net/rxrpc/ |
D | proc.c | 256 unsigned int bucket, n; in rxrpc_peer_seq_start() local 266 bucket = *_pos >> shift; in rxrpc_peer_seq_start() 268 if (bucket >= HASH_SIZE(rxnet->peer_hash)) { in rxrpc_peer_seq_start() 273 if (bucket == 0) in rxrpc_peer_seq_start() 279 p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1); in rxrpc_peer_seq_start() 282 bucket++; in rxrpc_peer_seq_start() 284 *_pos = (bucket << shift) | n; in rxrpc_peer_seq_start() 291 unsigned int bucket, n; in rxrpc_peer_seq_next() local 298 bucket = *_pos >> shift; in rxrpc_peer_seq_next() 300 p = seq_hlist_next_rcu(v, &rxnet->peer_hash[bucket], _pos); in rxrpc_peer_seq_next() [all …]
|
/Linux-v5.10/drivers/cpuidle/governors/ |
D | menu.c | 114 unsigned int bucket; member 122 int bucket = 0; in which_bucket() local 131 bucket = BUCKETS/2; in which_bucket() 134 return bucket; in which_bucket() 136 return bucket + 1; in which_bucket() 138 return bucket + 2; in which_bucket() 140 return bucket + 3; in which_bucket() 142 return bucket + 4; in which_bucket() 143 return bucket + 5; in which_bucket() 286 data->bucket = which_bucket(data->next_timer_ns, nr_iowaiters); in menu_select() [all …]
|
/Linux-v5.10/kernel/dma/ |
D | debug.c | 263 static void put_hash_bucket(struct hash_bucket *bucket, in put_hash_bucket() argument 265 __releases(&bucket->lock) in put_hash_bucket() 267 spin_unlock_irqrestore(&bucket->lock, flags); in put_hash_bucket() 292 static struct dma_debug_entry *__hash_bucket_find(struct hash_bucket *bucket, in __hash_bucket_find() argument 299 list_for_each_entry(entry, &bucket->list, list) { in __hash_bucket_find() 342 static struct dma_debug_entry *bucket_find_exact(struct hash_bucket *bucket, in bucket_find_exact() argument 345 return __hash_bucket_find(bucket, ref, exact_match); in bucket_find_exact() 348 static struct dma_debug_entry *bucket_find_contain(struct hash_bucket **bucket, in bucket_find_contain() argument 358 entry = __hash_bucket_find(*bucket, ref, containing_match); in bucket_find_contain() 366 put_hash_bucket(*bucket, *flags); in bucket_find_contain() [all …]
|
/Linux-v5.10/Documentation/userspace-api/media/v4l/ |
D | pixfmt-meta-vsp1-hgt.rst | 28 The Saturation position **n** (0 - 31) of the bucket in the matrix is 33 The Hue position **m** (0 - 5) of the bucket in the matrix depends on 101 - :cspan:`4` Histogram bucket (m=0, n=0) [31:0] 103 - :cspan:`4` Histogram bucket (m=0, n=1) [31:0] 107 - :cspan:`4` Histogram bucket (m=0, n=31) [31:0] 109 - :cspan:`4` Histogram bucket (m=1, n=0) [31:0] 113 - :cspan:`4` Histogram bucket (m=2, n=0) [31:0] 117 - :cspan:`4` Histogram bucket (m=3, n=0) [31:0] 121 - :cspan:`4` Histogram bucket (m=4, n=0) [31:0] 125 - :cspan:`4` Histogram bucket (m=5, n=0) [31:0] [all …]
|
/Linux-v5.10/fs/ocfs2/ |
D | xattr.c | 127 struct ocfs2_xattr_bucket *bucket; member 281 struct ocfs2_xattr_bucket *bucket, 303 struct ocfs2_xattr_bucket *bucket, 324 struct ocfs2_xattr_bucket *bucket; in ocfs2_xattr_bucket_new() local 329 bucket = kzalloc(sizeof(struct ocfs2_xattr_bucket), GFP_NOFS); in ocfs2_xattr_bucket_new() 330 if (bucket) { in ocfs2_xattr_bucket_new() 331 bucket->bu_inode = inode; in ocfs2_xattr_bucket_new() 332 bucket->bu_blocks = blks; in ocfs2_xattr_bucket_new() 335 return bucket; in ocfs2_xattr_bucket_new() 338 static void ocfs2_xattr_bucket_relse(struct ocfs2_xattr_bucket *bucket) in ocfs2_xattr_bucket_relse() argument [all …]
|
/Linux-v5.10/arch/mips/netlogic/xlr/ |
D | fmn.c | 72 int bucket, rv; in fmn_message_handler() local 86 for (bucket = 0; bucket < 8; bucket++) { in fmn_message_handler() 88 if (bkt_status & (1 << bucket)) in fmn_message_handler() 90 rv = nlm_fmn_receive(bucket, &size, &code, &src_stnid, in fmn_message_handler() 101 hndlr->action(bucket, src_stnid, size, code, in fmn_message_handler()
|
/Linux-v5.10/drivers/md/bcache/ |
D | alloc.c | 76 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) in bch_inc_gen() 89 struct bucket *b; in bch_rescale_priorities() 125 static inline bool can_inc_bucket_gen(struct bucket *b) in can_inc_bucket_gen() 130 bool bch_can_invalidate_bucket(struct cache *ca, struct bucket *b) in bch_can_invalidate_bucket() 140 void __bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) in __bch_invalidate_one_bucket() 153 static void bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) in bch_invalidate_one_bucket() 181 struct bucket *b; in invalidate_buckets_lru() 218 struct bucket *b; in invalidate_buckets_fifo() 241 struct bucket *b; in invalidate_buckets_random() 302 static int bch_allocator_push(struct cache *ca, long bucket) in bch_allocator_push() argument [all …]
|
/Linux-v5.10/drivers/md/ |
D | dm-clone-target.c | 570 #define bucket_lock_irqsave(bucket, flags) \ argument 571 spin_lock_irqsave(&(bucket)->lock, flags) 573 #define bucket_unlock_irqrestore(bucket, flags) \ argument 574 spin_unlock_irqrestore(&(bucket)->lock, flags) 576 #define bucket_lock_irq(bucket) \ argument 577 spin_lock_irq(&(bucket)->lock) 579 #define bucket_unlock_irq(bucket) \ argument 580 spin_unlock_irq(&(bucket)->lock) 585 struct hash_table_bucket *bucket; in hash_table_init() local 594 bucket = clone->ht + i; in hash_table_init() [all …]
|
/Linux-v5.10/lib/ |
D | stackdepot.c | 178 static inline struct stack_record *find_stack(struct stack_record *bucket, in find_stack() argument 184 for (found = bucket; found; found = found->next) { in find_stack() 239 struct stack_record *found = NULL, **bucket; in stack_depot_save() local 250 bucket = &stack_table[hash & STACK_HASH_MASK]; in stack_depot_save() 257 found = find_stack(smp_load_acquire(bucket), entries, in stack_depot_save() 286 found = find_stack(*bucket, entries, nr_entries, hash); in stack_depot_save() 292 new->next = *bucket; in stack_depot_save() 297 smp_store_release(bucket, new); in stack_depot_save()
|
/Linux-v5.10/include/trace/events/ |
D | bcache.h | 68 __field(size_t, bucket ) 72 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 75 TP_printk("bucket %zu", __entry->bucket) 267 __field(size_t, bucket ) 273 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 279 __entry->bucket, __entry->block, __entry->keys) 370 __field(size_t, bucket ) 375 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 379 TP_printk("bucket %zu keys %u", __entry->bucket, __entry->keys) 429 TP_PROTO(struct cache *ca, size_t bucket), [all …]
|
/Linux-v5.10/drivers/md/persistent-data/ |
D | dm-transaction-manager.c | 106 unsigned bucket = dm_hash_block(b, DM_HASH_MASK); in is_shadow() local 110 hlist_for_each_entry(si, tm->buckets + bucket, hlist) in is_shadow() 126 unsigned bucket; in insert_shadow() local 132 bucket = dm_hash_block(b, DM_HASH_MASK); in insert_shadow() 134 hlist_add_head(&si->hlist, tm->buckets + bucket); in insert_shadow() 143 struct hlist_head *bucket; in wipe_shadow_table() local 148 bucket = tm->buckets + i; in wipe_shadow_table() 149 hlist_for_each_entry_safe(si, tmp, bucket, hlist) in wipe_shadow_table() 152 INIT_HLIST_HEAD(bucket); in wipe_shadow_table()
|
/Linux-v5.10/arch/sparc/kernel/ |
D | irq_64.c | 206 struct ino_bucket bucket; member 257 struct ino_bucket *bucket; in cookie_exists() local 268 bucket = (struct ino_bucket *) __va(cookie); in cookie_exists() 269 irq = bucket->__irq; in cookie_exists() 278 struct ino_bucket *bucket; in sysino_exists() local 281 bucket = &ivector_table[sysino]; in sysino_exists() 282 irq = bucket_get_irq(__pa(bucket)); in sysino_exists() 615 struct ino_bucket *bucket; in build_irq() local 622 bucket = &ivector_table[ino]; in build_irq() 623 irq = bucket_get_irq(__pa(bucket)); in build_irq() [all …]
|