/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_icm_pool.c | 171 struct mlx5dr_icm_bucket *bucket = chunk->bucket; in dr_icm_chunk_ste_init() local 173 chunk->ste_arr = kvzalloc(bucket->num_of_entries * in dr_icm_chunk_ste_init() 178 chunk->hw_ste_arr = kvzalloc(bucket->num_of_entries * in dr_icm_chunk_ste_init() 183 chunk->miss_list = kvmalloc(bucket->num_of_entries * in dr_icm_chunk_ste_init() 197 static int dr_icm_chunks_create(struct mlx5dr_icm_bucket *bucket) in dr_icm_chunks_create() argument 200 struct mlx5dr_icm_pool *pool = bucket->pool; in dr_icm_chunks_create() 207 mr_req_size = bucket->num_of_entries * bucket->entry_size; in dr_icm_chunks_create() 246 chunk->bucket = bucket; in dr_icm_chunks_create() 252 chunk->num_of_entries = bucket->num_of_entries; in dr_icm_chunks_create() 253 chunk->byte_size = chunk->num_of_entries * bucket->entry_size; in dr_icm_chunks_create() [all …]
|
/Linux-v5.4/net/ceph/crush/ |
D | mapper.c | 74 static int bucket_perm_choose(const struct crush_bucket *bucket, in bucket_perm_choose() argument 78 unsigned int pr = r % bucket->size; in bucket_perm_choose() 83 dprintk("bucket %d new x=%d\n", bucket->id, x); in bucket_perm_choose() 88 s = crush_hash32_3(bucket->hash, x, bucket->id, 0) % in bucket_perm_choose() 89 bucket->size; in bucket_perm_choose() 95 for (i = 0; i < bucket->size; i++) in bucket_perm_choose() 100 for (i = 1; i < bucket->size; i++) in bucket_perm_choose() 112 if (p < bucket->size - 1) { in bucket_perm_choose() 113 i = crush_hash32_3(bucket->hash, x, bucket->id, p) % in bucket_perm_choose() 114 (bucket->size - p); in bucket_perm_choose() [all …]
|
/Linux-v5.4/block/ |
D | blk-stat.c | 56 int bucket; in blk_stat_add() local 68 bucket = cb->bucket_fn(rq); in blk_stat_add() 69 if (bucket < 0) in blk_stat_add() 72 stat = &get_cpu_ptr(cb->cpu_stat)[bucket]; in blk_stat_add() 82 unsigned int bucket; in blk_stat_timer_fn() local 85 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn() 86 blk_rq_stat_init(&cb->stat[bucket]); in blk_stat_timer_fn() 92 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn() 93 blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); in blk_stat_timer_fn() 94 blk_rq_stat_init(&cpu_stat[bucket]); in blk_stat_timer_fn() [all …]
|
D | kyber-iosched.c | 215 unsigned int bucket; in flush_latency_buckets() local 217 for (bucket = 0; bucket < KYBER_LATENCY_BUCKETS; bucket++) in flush_latency_buckets() 218 buckets[bucket] += atomic_xchg(&cpu_buckets[bucket], 0); in flush_latency_buckets() 230 unsigned int bucket, samples = 0, percentile_samples; in calculate_percentile() local 232 for (bucket = 0; bucket < KYBER_LATENCY_BUCKETS; bucket++) in calculate_percentile() 233 samples += buckets[bucket]; in calculate_percentile() 251 for (bucket = 0; bucket < KYBER_LATENCY_BUCKETS - 1; bucket++) { in calculate_percentile() 252 if (buckets[bucket] >= percentile_samples) in calculate_percentile() 254 percentile_samples -= buckets[bucket]; in calculate_percentile() 260 bucket + 1, 1 << KYBER_LATENCY_SHIFT, samples); in calculate_percentile() [all …]
|
/Linux-v5.4/net/sched/ |
D | sch_hhf.c | 329 static struct sk_buff *dequeue_head(struct wdrr_bucket *bucket) in dequeue_head() argument 331 struct sk_buff *skb = bucket->head; in dequeue_head() 333 bucket->head = skb->next; in dequeue_head() 339 static void bucket_add(struct wdrr_bucket *bucket, struct sk_buff *skb) in bucket_add() argument 341 if (bucket->head == NULL) in bucket_add() 342 bucket->head = skb; in bucket_add() 344 bucket->tail->next = skb; in bucket_add() 345 bucket->tail = skb; in bucket_add() 352 struct wdrr_bucket *bucket; in hhf_drop() local 355 bucket = &q->buckets[WDRR_BUCKET_FOR_HH]; in hhf_drop() [all …]
|
/Linux-v5.4/drivers/infiniband/sw/rdmavt/ |
D | trace_qp.h | 60 TP_PROTO(struct rvt_qp *qp, u32 bucket), 61 TP_ARGS(qp, bucket), 65 __field(u32, bucket) 70 __entry->bucket = bucket; 76 __entry->bucket 81 TP_PROTO(struct rvt_qp *qp, u32 bucket), 82 TP_ARGS(qp, bucket)); 85 TP_PROTO(struct rvt_qp *qp, u32 bucket), 86 TP_ARGS(qp, bucket));
|
/Linux-v5.4/net/9p/ |
D | error.c | 181 int bucket; in p9_error_init() local 184 for (bucket = 0; bucket < ERRHASHSZ; bucket++) in p9_error_init() 185 INIT_HLIST_HEAD(&hash_errmap[bucket]); in p9_error_init() 190 bucket = jhash(c->name, c->namelen, 0) % ERRHASHSZ; in p9_error_init() 192 hlist_add_head(&c->list, &hash_errmap[bucket]); in p9_error_init() 210 int bucket; in p9_errstr2errno() local 214 bucket = jhash(errstr, len, 0) % ERRHASHSZ; in p9_errstr2errno() 215 hlist_for_each_entry(c, &hash_errmap[bucket], list) { in p9_errstr2errno()
|
/Linux-v5.4/fs/dlm/ |
D | debug_fs.c | 368 unsigned bucket; member 427 unsigned bucket, entry; in table_seq_start() local 430 bucket = n >> 32; in table_seq_start() 433 if (bucket >= ls->ls_rsbtbl_size) in table_seq_start() 450 tree = toss ? &ls->ls_rsbtbl[bucket].toss : &ls->ls_rsbtbl[bucket].keep; in table_seq_start() 452 spin_lock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start() 459 ri->bucket = bucket; in table_seq_start() 460 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start() 465 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start() 475 bucket++; in table_seq_start() [all …]
|
/Linux-v5.4/net/vmw_vsock/ |
D | diag.c | 52 unsigned int bucket; in vsock_diag_dump() local 63 bucket = cb->args[1]; in vsock_diag_dump() 72 while (bucket < ARRAY_SIZE(vsock_bind_table)) { in vsock_diag_dump() 73 struct list_head *head = &vsock_bind_table[bucket]; in vsock_diag_dump() 94 bucket++; in vsock_diag_dump() 98 bucket = 0; in vsock_diag_dump() 102 while (bucket < ARRAY_SIZE(vsock_connected_table)) { in vsock_diag_dump() 103 struct list_head *head = &vsock_connected_table[bucket]; in vsock_diag_dump() 128 bucket++; in vsock_diag_dump() 135 cb->args[1] = bucket; in vsock_diag_dump()
|
/Linux-v5.4/net/rxrpc/ |
D | proc.c | 256 unsigned int bucket, n; in rxrpc_peer_seq_start() local 266 bucket = *_pos >> shift; in rxrpc_peer_seq_start() 268 if (bucket >= HASH_SIZE(rxnet->peer_hash)) { in rxrpc_peer_seq_start() 273 if (bucket == 0) in rxrpc_peer_seq_start() 279 p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1); in rxrpc_peer_seq_start() 282 bucket++; in rxrpc_peer_seq_start() 284 *_pos = (bucket << shift) | n; in rxrpc_peer_seq_start() 291 unsigned int bucket, n; in rxrpc_peer_seq_next() local 298 bucket = *_pos >> shift; in rxrpc_peer_seq_next() 300 p = seq_hlist_next_rcu(v, &rxnet->peer_hash[bucket], _pos); in rxrpc_peer_seq_next() [all …]
|
/Linux-v5.4/drivers/cpuidle/governors/ |
D | menu.c | 124 unsigned int bucket; member 132 int bucket = 0; in which_bucket() local 141 bucket = BUCKETS/2; in which_bucket() 144 return bucket; in which_bucket() 146 return bucket + 1; in which_bucket() 148 return bucket + 2; in which_bucket() 150 return bucket + 3; in which_bucket() 152 return bucket + 4; in which_bucket() 153 return bucket + 5; in which_bucket() 296 data->bucket = which_bucket(data->next_timer_us, nr_iowaiters); in menu_select() [all …]
|
/Linux-v5.4/kernel/dma/ |
D | debug.c | 257 static void put_hash_bucket(struct hash_bucket *bucket, in put_hash_bucket() argument 259 __releases(&bucket->lock) in put_hash_bucket() 263 spin_unlock_irqrestore(&bucket->lock, __flags); in put_hash_bucket() 288 static struct dma_debug_entry *__hash_bucket_find(struct hash_bucket *bucket, in __hash_bucket_find() argument 295 list_for_each_entry(entry, &bucket->list, list) { in __hash_bucket_find() 338 static struct dma_debug_entry *bucket_find_exact(struct hash_bucket *bucket, in bucket_find_exact() argument 341 return __hash_bucket_find(bucket, ref, exact_match); in bucket_find_exact() 344 static struct dma_debug_entry *bucket_find_contain(struct hash_bucket **bucket, in bucket_find_contain() argument 354 entry = __hash_bucket_find(*bucket, ref, containing_match); in bucket_find_contain() 362 put_hash_bucket(*bucket, flags); in bucket_find_contain() [all …]
|
/Linux-v5.4/fs/nfs/ |
D | pnfs_nfs.c | 77 struct pnfs_commit_bucket *bucket; in pnfs_generic_clear_request_commit() local 79 bucket = list_first_entry(&req->wb_list, in pnfs_generic_clear_request_commit() 82 freeme = bucket->wlseg; in pnfs_generic_clear_request_commit() 83 bucket->wlseg = NULL; in pnfs_generic_clear_request_commit() 92 pnfs_generic_scan_ds_commit_list(struct pnfs_commit_bucket *bucket, in pnfs_generic_scan_ds_commit_list() argument 96 struct list_head *src = &bucket->written; in pnfs_generic_scan_ds_commit_list() 97 struct list_head *dst = &bucket->committing; in pnfs_generic_scan_ds_commit_list() 105 if (bucket->clseg == NULL) in pnfs_generic_scan_ds_commit_list() 106 bucket->clseg = pnfs_get_lseg(bucket->wlseg); in pnfs_generic_scan_ds_commit_list() 108 pnfs_put_lseg(bucket->wlseg); in pnfs_generic_scan_ds_commit_list() [all …]
|
/Linux-v5.4/fs/ocfs2/ |
D | xattr.c | 127 struct ocfs2_xattr_bucket *bucket; member 281 struct ocfs2_xattr_bucket *bucket, 303 struct ocfs2_xattr_bucket *bucket, 324 struct ocfs2_xattr_bucket *bucket; in ocfs2_xattr_bucket_new() local 329 bucket = kzalloc(sizeof(struct ocfs2_xattr_bucket), GFP_NOFS); in ocfs2_xattr_bucket_new() 330 if (bucket) { in ocfs2_xattr_bucket_new() 331 bucket->bu_inode = inode; in ocfs2_xattr_bucket_new() 332 bucket->bu_blocks = blks; in ocfs2_xattr_bucket_new() 335 return bucket; in ocfs2_xattr_bucket_new() 338 static void ocfs2_xattr_bucket_relse(struct ocfs2_xattr_bucket *bucket) in ocfs2_xattr_bucket_relse() argument [all …]
|
/Linux-v5.4/arch/mips/netlogic/xlr/ |
D | fmn.c | 72 int bucket, rv; in fmn_message_handler() local 86 for (bucket = 0; bucket < 8; bucket++) { in fmn_message_handler() 88 if (bkt_status & (1 << bucket)) in fmn_message_handler() 90 rv = nlm_fmn_receive(bucket, &size, &code, &src_stnid, in fmn_message_handler() 101 hndlr->action(bucket, src_stnid, size, code, in fmn_message_handler()
|
/Linux-v5.4/Documentation/media/uapi/v4l/ |
D | pixfmt-meta-vsp1-hgt.rst | 35 The Saturation position **n** (0 - 31) of the bucket in the matrix is 40 The Hue position **m** (0 - 5) of the bucket in the matrix depends on 108 - :cspan:`4` Histogram bucket (m=0, n=0) [31:0] 110 - :cspan:`4` Histogram bucket (m=0, n=1) [31:0] 114 - :cspan:`4` Histogram bucket (m=0, n=31) [31:0] 116 - :cspan:`4` Histogram bucket (m=1, n=0) [31:0] 120 - :cspan:`4` Histogram bucket (m=2, n=0) [31:0] 124 - :cspan:`4` Histogram bucket (m=3, n=0) [31:0] 128 - :cspan:`4` Histogram bucket (m=4, n=0) [31:0] 132 - :cspan:`4` Histogram bucket (m=5, n=0) [31:0] [all …]
|
/Linux-v5.4/lib/ |
D | stackdepot.c | 173 static inline struct stack_record *find_stack(struct stack_record *bucket, in find_stack() argument 179 for (found = bucket; found; found = found->next) { in find_stack() 223 struct stack_record *found = NULL, **bucket; in stack_depot_save() local 234 bucket = &stack_table[hash & STACK_HASH_MASK]; in stack_depot_save() 241 found = find_stack(smp_load_acquire(bucket), entries, in stack_depot_save() 270 found = find_stack(*bucket, entries, nr_entries, hash); in stack_depot_save() 276 new->next = *bucket; in stack_depot_save() 281 smp_store_release(bucket, new); in stack_depot_save()
|
/Linux-v5.4/drivers/md/bcache/ |
D | alloc.c | 76 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) in bch_inc_gen() 89 struct bucket *b; in bch_rescale_priorities() 126 static inline bool can_inc_bucket_gen(struct bucket *b) in can_inc_bucket_gen() 131 bool bch_can_invalidate_bucket(struct cache *ca, struct bucket *b) in bch_can_invalidate_bucket() 141 void __bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) in __bch_invalidate_one_bucket() 154 static void bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) in bch_invalidate_one_bucket() 182 struct bucket *b; in invalidate_buckets_lru() 219 struct bucket *b; in invalidate_buckets_fifo() 242 struct bucket *b; in invalidate_buckets_random() 303 static int bch_allocator_push(struct cache *ca, long bucket) in bch_allocator_push() argument [all …]
|
/Linux-v5.4/include/trace/events/ |
D | bcache.h | 68 __field(size_t, bucket ) 72 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 75 TP_printk("bucket %zu", __entry->bucket) 267 __field(size_t, bucket ) 273 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 278 TP_printk("bucket %zu", __entry->bucket) 369 __field(size_t, bucket ) 374 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0); 378 TP_printk("bucket %zu keys %u", __entry->bucket, __entry->keys) 428 TP_PROTO(struct cache *ca, size_t bucket), [all …]
|
/Linux-v5.4/drivers/md/ |
D | dm-clone-target.c | 551 #define bucket_lock_irqsave(bucket, flags) \ argument 552 spin_lock_irqsave(&(bucket)->lock, flags) 554 #define bucket_unlock_irqrestore(bucket, flags) \ argument 555 spin_unlock_irqrestore(&(bucket)->lock, flags) 560 struct hash_table_bucket *bucket; in hash_table_init() local 569 bucket = clone->ht + i; in hash_table_init() 571 INIT_HLIST_HEAD(&bucket->head); in hash_table_init() 572 spin_lock_init(&bucket->lock); in hash_table_init() 594 static struct dm_clone_region_hydration *__hash_find(struct hash_table_bucket *bucket, in __hash_find() argument 599 hlist_for_each_entry(hd, &bucket->head, h) { in __hash_find() [all …]
|
/Linux-v5.4/kernel/bpf/ |
D | stackmap.c | 350 struct stack_map_bucket *bucket, *new_bucket, *old_bucket; in BPF_CALL_3() local 386 bucket = READ_ONCE(smap->buckets[id]); in BPF_CALL_3() 388 hash_matches = bucket && bucket->hash == hash; in BPF_CALL_3() 404 if (hash_matches && bucket->nr == trace_nr && in BPF_CALL_3() 405 memcmp(bucket->data, new_bucket->data, trace_len) == 0) { in BPF_CALL_3() 409 if (bucket && !(flags & BPF_F_REUSE_STACKID)) { in BPF_CALL_3() 414 if (hash_matches && bucket->nr == trace_nr && in BPF_CALL_3() 415 memcmp(bucket->data, ips, trace_len) == 0) in BPF_CALL_3() 417 if (bucket && !(flags & BPF_F_REUSE_STACKID)) in BPF_CALL_3() 522 struct stack_map_bucket *bucket, *old_bucket; in bpf_stackmap_copy() local [all …]
|
/Linux-v5.4/net/core/ |
D | sock_map.c | 562 struct bpf_htab_bucket *bucket; in __sock_hash_lookup_elem() local 568 bucket = sock_hash_select_bucket(htab, hash); in __sock_hash_lookup_elem() 569 elem = sock_hash_lookup_elem_raw(&bucket->head, hash, key, key_size); in __sock_hash_lookup_elem() 586 struct bpf_htab_bucket *bucket; in sock_hash_delete_from_link() local 589 bucket = sock_hash_select_bucket(htab, elem->hash); in sock_hash_delete_from_link() 595 raw_spin_lock_bh(&bucket->lock); in sock_hash_delete_from_link() 596 elem_probe = sock_hash_lookup_elem_raw(&bucket->head, elem->hash, in sock_hash_delete_from_link() 603 raw_spin_unlock_bh(&bucket->lock); in sock_hash_delete_from_link() 610 struct bpf_htab_bucket *bucket; in sock_hash_delete_elem() local 615 bucket = sock_hash_select_bucket(htab, hash); in sock_hash_delete_elem() [all …]
|
/Linux-v5.4/drivers/md/persistent-data/ |
D | dm-transaction-manager.c | 106 unsigned bucket = dm_hash_block(b, DM_HASH_MASK); in is_shadow() local 110 hlist_for_each_entry(si, tm->buckets + bucket, hlist) in is_shadow() 126 unsigned bucket; in insert_shadow() local 132 bucket = dm_hash_block(b, DM_HASH_MASK); in insert_shadow() 134 hlist_add_head(&si->hlist, tm->buckets + bucket); in insert_shadow() 143 struct hlist_head *bucket; in wipe_shadow_table() local 148 bucket = tm->buckets + i; in wipe_shadow_table() 149 hlist_for_each_entry_safe(si, tmp, bucket, hlist) in wipe_shadow_table() 152 INIT_HLIST_HEAD(bucket); in wipe_shadow_table()
|
/Linux-v5.4/arch/sparc/kernel/ |
D | irq_64.c | 206 struct ino_bucket bucket; member 257 struct ino_bucket *bucket; in cookie_exists() local 268 bucket = (struct ino_bucket *) __va(cookie); in cookie_exists() 269 irq = bucket->__irq; in cookie_exists() 278 struct ino_bucket *bucket; in sysino_exists() local 281 bucket = &ivector_table[sysino]; in sysino_exists() 282 irq = bucket_get_irq(__pa(bucket)); in sysino_exists() 615 struct ino_bucket *bucket; in build_irq() local 622 bucket = &ivector_table[ino]; in build_irq() 623 irq = bucket_get_irq(__pa(bucket)); in build_irq() [all …]
|
/Linux-v5.4/net/atm/ |
D | proc.c | 69 int bucket; member 78 static int __vcc_walk(struct sock **sock, int family, int *bucket, loff_t l) in __vcc_walk() argument 83 for (*bucket = 0; *bucket < VCC_HTABLE_SIZE; ++*bucket) { in __vcc_walk() 84 struct hlist_head *head = &vcc_hash[*bucket]; in __vcc_walk() 98 if (!sk && ++*bucket < VCC_HTABLE_SIZE) { in __vcc_walk() 99 sk = sk_head(&vcc_hash[*bucket]); in __vcc_walk() 113 return __vcc_walk(&state->sk, family, &state->bucket, l) ? in vcc_walk()
|