Home
last modified time | relevance | path

Searched refs:bucket (Results 1 – 25 of 166) sorted by relevance

1234567

/Linux-v5.10/drivers/net/ethernet/mellanox/mlx5/core/steering/
Ddr_icm_pool.c181 struct mlx5dr_icm_bucket *bucket = chunk->bucket; in dr_icm_chunk_ste_init() local
183 chunk->ste_arr = kvzalloc(bucket->num_of_entries * in dr_icm_chunk_ste_init()
188 chunk->hw_ste_arr = kvzalloc(bucket->num_of_entries * in dr_icm_chunk_ste_init()
193 chunk->miss_list = kvmalloc(bucket->num_of_entries * in dr_icm_chunk_ste_init()
207 static int dr_icm_chunks_create(struct mlx5dr_icm_bucket *bucket) in dr_icm_chunks_create() argument
210 struct mlx5dr_icm_pool *pool = bucket->pool; in dr_icm_chunks_create()
215 mr_req_size = bucket->num_of_entries * bucket->entry_size; in dr_icm_chunks_create()
243 chunk->bucket = bucket; in dr_icm_chunks_create()
249 chunk->num_of_entries = bucket->num_of_entries; in dr_icm_chunks_create()
250 chunk->byte_size = chunk->num_of_entries * bucket->entry_size; in dr_icm_chunks_create()
[all …]
/Linux-v5.10/net/mptcp/
Dtoken.c112 struct token_bucket *bucket; in mptcp_token_new_request() local
123 bucket = token_bucket(token); in mptcp_token_new_request()
124 spin_lock_bh(&bucket->lock); in mptcp_token_new_request()
125 if (__token_bucket_busy(bucket, token)) { in mptcp_token_new_request()
126 spin_unlock_bh(&bucket->lock); in mptcp_token_new_request()
130 hlist_nulls_add_head_rcu(&subflow_req->token_node, &bucket->req_chain); in mptcp_token_new_request()
131 bucket->chain_len++; in mptcp_token_new_request()
132 spin_unlock_bh(&bucket->lock); in mptcp_token_new_request()
157 struct token_bucket *bucket; in mptcp_token_new_connect() local
166 bucket = token_bucket(subflow->token); in mptcp_token_new_connect()
[all …]
/Linux-v5.10/net/ceph/crush/
Dmapper.c74 static int bucket_perm_choose(const struct crush_bucket *bucket, in bucket_perm_choose() argument
78 unsigned int pr = r % bucket->size; in bucket_perm_choose()
83 dprintk("bucket %d new x=%d\n", bucket->id, x); in bucket_perm_choose()
88 s = crush_hash32_3(bucket->hash, x, bucket->id, 0) % in bucket_perm_choose()
89 bucket->size; in bucket_perm_choose()
95 for (i = 0; i < bucket->size; i++) in bucket_perm_choose()
100 for (i = 1; i < bucket->size; i++) in bucket_perm_choose()
112 if (p < bucket->size - 1) { in bucket_perm_choose()
113 i = crush_hash32_3(bucket->hash, x, bucket->id, p) % in bucket_perm_choose()
114 (bucket->size - p); in bucket_perm_choose()
[all …]
/Linux-v5.10/block/
Dblk-stat.c56 int bucket, cpu; in blk_stat_add() local
69 bucket = cb->bucket_fn(rq); in blk_stat_add()
70 if (bucket < 0) in blk_stat_add()
73 stat = &per_cpu_ptr(cb->cpu_stat, cpu)[bucket]; in blk_stat_add()
83 unsigned int bucket; in blk_stat_timer_fn() local
86 for (bucket = 0; bucket < cb->buckets; bucket++) in blk_stat_timer_fn()
87 blk_rq_stat_init(&cb->stat[bucket]); in blk_stat_timer_fn()
93 for (bucket = 0; bucket < cb->buckets; bucket++) { in blk_stat_timer_fn()
94 blk_rq_stat_sum(&cb->stat[bucket], &cpu_stat[bucket]); in blk_stat_timer_fn()
95 blk_rq_stat_init(&cpu_stat[bucket]); in blk_stat_timer_fn()
[all …]
Dkyber-iosched.c215 unsigned int bucket; in flush_latency_buckets() local
217 for (bucket = 0; bucket < KYBER_LATENCY_BUCKETS; bucket++) in flush_latency_buckets()
218 buckets[bucket] += atomic_xchg(&cpu_buckets[bucket], 0); in flush_latency_buckets()
230 unsigned int bucket, samples = 0, percentile_samples; in calculate_percentile() local
232 for (bucket = 0; bucket < KYBER_LATENCY_BUCKETS; bucket++) in calculate_percentile()
233 samples += buckets[bucket]; in calculate_percentile()
251 for (bucket = 0; bucket < KYBER_LATENCY_BUCKETS - 1; bucket++) { in calculate_percentile()
252 if (buckets[bucket] >= percentile_samples) in calculate_percentile()
254 percentile_samples -= buckets[bucket]; in calculate_percentile()
260 bucket + 1, 1 << KYBER_LATENCY_SHIFT, samples); in calculate_percentile()
[all …]
/Linux-v5.10/net/sched/
Dsch_hhf.c329 static struct sk_buff *dequeue_head(struct wdrr_bucket *bucket) in dequeue_head() argument
331 struct sk_buff *skb = bucket->head; in dequeue_head()
333 bucket->head = skb->next; in dequeue_head()
339 static void bucket_add(struct wdrr_bucket *bucket, struct sk_buff *skb) in bucket_add() argument
341 if (bucket->head == NULL) in bucket_add()
342 bucket->head = skb; in bucket_add()
344 bucket->tail->next = skb; in bucket_add()
345 bucket->tail = skb; in bucket_add()
352 struct wdrr_bucket *bucket; in hhf_drop() local
355 bucket = &q->buckets[WDRR_BUCKET_FOR_HH]; in hhf_drop()
[all …]
/Linux-v5.10/drivers/infiniband/sw/rdmavt/
Dtrace_qp.h60 TP_PROTO(struct rvt_qp *qp, u32 bucket),
61 TP_ARGS(qp, bucket),
65 __field(u32, bucket)
70 __entry->bucket = bucket;
76 __entry->bucket
81 TP_PROTO(struct rvt_qp *qp, u32 bucket),
82 TP_ARGS(qp, bucket));
85 TP_PROTO(struct rvt_qp *qp, u32 bucket),
86 TP_ARGS(qp, bucket));
/Linux-v5.10/drivers/interconnect/qcom/
Dbcm-voter.c71 size_t i, bucket; in bcm_aggregate() local
76 for (bucket = 0; bucket < QCOM_ICC_NUM_BUCKETS; bucket++) { in bcm_aggregate()
79 temp = bcm_div(node->sum_avg[bucket] * bcm->aux_data.width, in bcm_aggregate()
81 agg_avg[bucket] = max(agg_avg[bucket], temp); in bcm_aggregate()
83 temp = bcm_div(node->max_peak[bucket] * bcm->aux_data.width, in bcm_aggregate()
85 agg_peak[bucket] = max(agg_peak[bucket], temp); in bcm_aggregate()
88 temp = agg_avg[bucket] * bcm->vote_scale; in bcm_aggregate()
89 bcm->vote_x[bucket] = bcm_div(temp, bcm->aux_data.unit); in bcm_aggregate()
91 temp = agg_peak[bucket] * bcm->vote_scale; in bcm_aggregate()
92 bcm->vote_y[bucket] = bcm_div(temp, bcm->aux_data.unit); in bcm_aggregate()
[all …]
/Linux-v5.10/fs/nfs/
Dnfs42xattr.c87 struct nfs4_xattr_bucket *bucket; member
238 entry->bucket = NULL; in nfs4_xattr_alloc_entry()
389 struct nfs4_xattr_bucket *bucket; in nfs4_xattr_discard_cache() local
395 bucket = &cache->buckets[i]; in nfs4_xattr_discard_cache()
397 spin_lock(&bucket->lock); in nfs4_xattr_discard_cache()
398 bucket->draining = true; in nfs4_xattr_discard_cache()
399 hlist_for_each_entry_safe(entry, n, &bucket->hlist, hnode) { in nfs4_xattr_discard_cache()
404 spin_unlock(&bucket->lock); in nfs4_xattr_discard_cache()
512 nfs4_xattr_get_entry(struct nfs4_xattr_bucket *bucket, const char *name) in nfs4_xattr_get_entry() argument
518 hlist_for_each_entry(entry, &bucket->hlist, hnode) { in nfs4_xattr_get_entry()
[all …]
Dpnfs_nfs.c63 pnfs_free_bucket_lseg(struct pnfs_commit_bucket *bucket) in pnfs_free_bucket_lseg() argument
65 if (list_empty(&bucket->committing) && list_empty(&bucket->written)) { in pnfs_free_bucket_lseg()
66 struct pnfs_layout_segment *freeme = bucket->lseg; in pnfs_free_bucket_lseg()
67 bucket->lseg = NULL; in pnfs_free_bucket_lseg()
87 struct pnfs_commit_bucket *bucket; in pnfs_generic_clear_request_commit() local
89 bucket = list_first_entry(&req->wb_list, in pnfs_generic_clear_request_commit()
92 freeme = pnfs_free_bucket_lseg(bucket); in pnfs_generic_clear_request_commit()
245 pnfs_bucket_scan_ds_commit_list(struct pnfs_commit_bucket *bucket, in pnfs_bucket_scan_ds_commit_list() argument
249 struct list_head *src = &bucket->written; in pnfs_bucket_scan_ds_commit_list()
250 struct list_head *dst = &bucket->committing; in pnfs_bucket_scan_ds_commit_list()
[all …]
/Linux-v5.10/net/9p/
Derror.c181 int bucket; in p9_error_init() local
184 for (bucket = 0; bucket < ERRHASHSZ; bucket++) in p9_error_init()
185 INIT_HLIST_HEAD(&hash_errmap[bucket]); in p9_error_init()
190 bucket = jhash(c->name, c->namelen, 0) % ERRHASHSZ; in p9_error_init()
192 hlist_add_head(&c->list, &hash_errmap[bucket]); in p9_error_init()
210 int bucket; in p9_errstr2errno() local
214 bucket = jhash(errstr, len, 0) % ERRHASHSZ; in p9_errstr2errno()
215 hlist_for_each_entry(c, &hash_errmap[bucket], list) { in p9_errstr2errno()
/Linux-v5.10/fs/dlm/
Ddebug_fs.c368 unsigned bucket; member
427 unsigned bucket, entry; in table_seq_start() local
430 bucket = n >> 32; in table_seq_start()
433 if (bucket >= ls->ls_rsbtbl_size) in table_seq_start()
450 tree = toss ? &ls->ls_rsbtbl[bucket].toss : &ls->ls_rsbtbl[bucket].keep; in table_seq_start()
452 spin_lock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start()
459 ri->bucket = bucket; in table_seq_start()
460 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start()
465 spin_unlock(&ls->ls_rsbtbl[bucket].lock); in table_seq_start()
475 bucket++; in table_seq_start()
[all …]
/Linux-v5.10/net/vmw_vsock/
Ddiag.c52 unsigned int bucket; in vsock_diag_dump() local
63 bucket = cb->args[1]; in vsock_diag_dump()
72 while (bucket < ARRAY_SIZE(vsock_bind_table)) { in vsock_diag_dump()
73 struct list_head *head = &vsock_bind_table[bucket]; in vsock_diag_dump()
94 bucket++; in vsock_diag_dump()
98 bucket = 0; in vsock_diag_dump()
102 while (bucket < ARRAY_SIZE(vsock_connected_table)) { in vsock_diag_dump()
103 struct list_head *head = &vsock_connected_table[bucket]; in vsock_diag_dump()
128 bucket++; in vsock_diag_dump()
135 cb->args[1] = bucket; in vsock_diag_dump()
/Linux-v5.10/net/rxrpc/
Dproc.c256 unsigned int bucket, n; in rxrpc_peer_seq_start() local
266 bucket = *_pos >> shift; in rxrpc_peer_seq_start()
268 if (bucket >= HASH_SIZE(rxnet->peer_hash)) { in rxrpc_peer_seq_start()
273 if (bucket == 0) in rxrpc_peer_seq_start()
279 p = seq_hlist_start_rcu(&rxnet->peer_hash[bucket], n - 1); in rxrpc_peer_seq_start()
282 bucket++; in rxrpc_peer_seq_start()
284 *_pos = (bucket << shift) | n; in rxrpc_peer_seq_start()
291 unsigned int bucket, n; in rxrpc_peer_seq_next() local
298 bucket = *_pos >> shift; in rxrpc_peer_seq_next()
300 p = seq_hlist_next_rcu(v, &rxnet->peer_hash[bucket], _pos); in rxrpc_peer_seq_next()
[all …]
/Linux-v5.10/drivers/cpuidle/governors/
Dmenu.c114 unsigned int bucket; member
122 int bucket = 0; in which_bucket() local
131 bucket = BUCKETS/2; in which_bucket()
134 return bucket; in which_bucket()
136 return bucket + 1; in which_bucket()
138 return bucket + 2; in which_bucket()
140 return bucket + 3; in which_bucket()
142 return bucket + 4; in which_bucket()
143 return bucket + 5; in which_bucket()
286 data->bucket = which_bucket(data->next_timer_ns, nr_iowaiters); in menu_select()
[all …]
/Linux-v5.10/kernel/dma/
Ddebug.c263 static void put_hash_bucket(struct hash_bucket *bucket, in put_hash_bucket() argument
265 __releases(&bucket->lock) in put_hash_bucket()
267 spin_unlock_irqrestore(&bucket->lock, flags); in put_hash_bucket()
292 static struct dma_debug_entry *__hash_bucket_find(struct hash_bucket *bucket, in __hash_bucket_find() argument
299 list_for_each_entry(entry, &bucket->list, list) { in __hash_bucket_find()
342 static struct dma_debug_entry *bucket_find_exact(struct hash_bucket *bucket, in bucket_find_exact() argument
345 return __hash_bucket_find(bucket, ref, exact_match); in bucket_find_exact()
348 static struct dma_debug_entry *bucket_find_contain(struct hash_bucket **bucket, in bucket_find_contain() argument
358 entry = __hash_bucket_find(*bucket, ref, containing_match); in bucket_find_contain()
366 put_hash_bucket(*bucket, *flags); in bucket_find_contain()
[all …]
/Linux-v5.10/Documentation/userspace-api/media/v4l/
Dpixfmt-meta-vsp1-hgt.rst28 The Saturation position **n** (0 - 31) of the bucket in the matrix is
33 The Hue position **m** (0 - 5) of the bucket in the matrix depends on
101 - :cspan:`4` Histogram bucket (m=0, n=0) [31:0]
103 - :cspan:`4` Histogram bucket (m=0, n=1) [31:0]
107 - :cspan:`4` Histogram bucket (m=0, n=31) [31:0]
109 - :cspan:`4` Histogram bucket (m=1, n=0) [31:0]
113 - :cspan:`4` Histogram bucket (m=2, n=0) [31:0]
117 - :cspan:`4` Histogram bucket (m=3, n=0) [31:0]
121 - :cspan:`4` Histogram bucket (m=4, n=0) [31:0]
125 - :cspan:`4` Histogram bucket (m=5, n=0) [31:0]
[all …]
/Linux-v5.10/fs/ocfs2/
Dxattr.c127 struct ocfs2_xattr_bucket *bucket; member
281 struct ocfs2_xattr_bucket *bucket,
303 struct ocfs2_xattr_bucket *bucket,
324 struct ocfs2_xattr_bucket *bucket; in ocfs2_xattr_bucket_new() local
329 bucket = kzalloc(sizeof(struct ocfs2_xattr_bucket), GFP_NOFS); in ocfs2_xattr_bucket_new()
330 if (bucket) { in ocfs2_xattr_bucket_new()
331 bucket->bu_inode = inode; in ocfs2_xattr_bucket_new()
332 bucket->bu_blocks = blks; in ocfs2_xattr_bucket_new()
335 return bucket; in ocfs2_xattr_bucket_new()
338 static void ocfs2_xattr_bucket_relse(struct ocfs2_xattr_bucket *bucket) in ocfs2_xattr_bucket_relse() argument
[all …]
/Linux-v5.10/arch/mips/netlogic/xlr/
Dfmn.c72 int bucket, rv; in fmn_message_handler() local
86 for (bucket = 0; bucket < 8; bucket++) { in fmn_message_handler()
88 if (bkt_status & (1 << bucket)) in fmn_message_handler()
90 rv = nlm_fmn_receive(bucket, &size, &code, &src_stnid, in fmn_message_handler()
101 hndlr->action(bucket, src_stnid, size, code, in fmn_message_handler()
/Linux-v5.10/drivers/md/bcache/
Dalloc.c76 uint8_t bch_inc_gen(struct cache *ca, struct bucket *b) in bch_inc_gen()
89 struct bucket *b; in bch_rescale_priorities()
125 static inline bool can_inc_bucket_gen(struct bucket *b) in can_inc_bucket_gen()
130 bool bch_can_invalidate_bucket(struct cache *ca, struct bucket *b) in bch_can_invalidate_bucket()
140 void __bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) in __bch_invalidate_one_bucket()
153 static void bch_invalidate_one_bucket(struct cache *ca, struct bucket *b) in bch_invalidate_one_bucket()
181 struct bucket *b; in invalidate_buckets_lru()
218 struct bucket *b; in invalidate_buckets_fifo()
241 struct bucket *b; in invalidate_buckets_random()
302 static int bch_allocator_push(struct cache *ca, long bucket) in bch_allocator_push() argument
[all …]
/Linux-v5.10/drivers/md/
Ddm-clone-target.c570 #define bucket_lock_irqsave(bucket, flags) \ argument
571 spin_lock_irqsave(&(bucket)->lock, flags)
573 #define bucket_unlock_irqrestore(bucket, flags) \ argument
574 spin_unlock_irqrestore(&(bucket)->lock, flags)
576 #define bucket_lock_irq(bucket) \ argument
577 spin_lock_irq(&(bucket)->lock)
579 #define bucket_unlock_irq(bucket) \ argument
580 spin_unlock_irq(&(bucket)->lock)
585 struct hash_table_bucket *bucket; in hash_table_init() local
594 bucket = clone->ht + i; in hash_table_init()
[all …]
/Linux-v5.10/lib/
Dstackdepot.c178 static inline struct stack_record *find_stack(struct stack_record *bucket, in find_stack() argument
184 for (found = bucket; found; found = found->next) { in find_stack()
239 struct stack_record *found = NULL, **bucket; in stack_depot_save() local
250 bucket = &stack_table[hash & STACK_HASH_MASK]; in stack_depot_save()
257 found = find_stack(smp_load_acquire(bucket), entries, in stack_depot_save()
286 found = find_stack(*bucket, entries, nr_entries, hash); in stack_depot_save()
292 new->next = *bucket; in stack_depot_save()
297 smp_store_release(bucket, new); in stack_depot_save()
/Linux-v5.10/include/trace/events/
Dbcache.h68 __field(size_t, bucket )
72 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0);
75 TP_printk("bucket %zu", __entry->bucket)
267 __field(size_t, bucket )
273 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0);
279 __entry->bucket, __entry->block, __entry->keys)
370 __field(size_t, bucket )
375 __entry->bucket = PTR_BUCKET_NR(b->c, &b->key, 0);
379 TP_printk("bucket %zu keys %u", __entry->bucket, __entry->keys)
429 TP_PROTO(struct cache *ca, size_t bucket),
[all …]
/Linux-v5.10/drivers/md/persistent-data/
Ddm-transaction-manager.c106 unsigned bucket = dm_hash_block(b, DM_HASH_MASK); in is_shadow() local
110 hlist_for_each_entry(si, tm->buckets + bucket, hlist) in is_shadow()
126 unsigned bucket; in insert_shadow() local
132 bucket = dm_hash_block(b, DM_HASH_MASK); in insert_shadow()
134 hlist_add_head(&si->hlist, tm->buckets + bucket); in insert_shadow()
143 struct hlist_head *bucket; in wipe_shadow_table() local
148 bucket = tm->buckets + i; in wipe_shadow_table()
149 hlist_for_each_entry_safe(si, tmp, bucket, hlist) in wipe_shadow_table()
152 INIT_HLIST_HEAD(bucket); in wipe_shadow_table()
/Linux-v5.10/arch/sparc/kernel/
Dirq_64.c206 struct ino_bucket bucket; member
257 struct ino_bucket *bucket; in cookie_exists() local
268 bucket = (struct ino_bucket *) __va(cookie); in cookie_exists()
269 irq = bucket->__irq; in cookie_exists()
278 struct ino_bucket *bucket; in sysino_exists() local
281 bucket = &ivector_table[sysino]; in sysino_exists()
282 irq = bucket_get_irq(__pa(bucket)); in sysino_exists()
615 struct ino_bucket *bucket; in build_irq() local
622 bucket = &ivector_table[ino]; in build_irq()
623 irq = bucket_get_irq(__pa(bucket)); in build_irq()
[all …]

1234567