/Linux-v6.6/lib/ |
D | sbitmap.c | 384 static unsigned int sbq_calc_wake_batch(struct sbitmap_queue *sbq, in sbq_calc_wake_batch() argument 406 shallow_depth = min(1U << sbq->sb.shift, sbq->min_shallow_depth); in sbq_calc_wake_batch() 407 depth = ((depth >> sbq->sb.shift) * shallow_depth + in sbq_calc_wake_batch() 408 min(depth & ((1U << sbq->sb.shift) - 1), shallow_depth)); in sbq_calc_wake_batch() 415 int sbitmap_queue_init_node(struct sbitmap_queue *sbq, unsigned int depth, in sbitmap_queue_init_node() argument 421 ret = sbitmap_init_node(&sbq->sb, depth, shift, flags, node, in sbitmap_queue_init_node() 426 sbq->min_shallow_depth = UINT_MAX; in sbitmap_queue_init_node() 427 sbq->wake_batch = sbq_calc_wake_batch(sbq, depth); in sbitmap_queue_init_node() 428 atomic_set(&sbq->wake_index, 0); in sbitmap_queue_init_node() 429 atomic_set(&sbq->ws_active, 0); in sbitmap_queue_init_node() [all …]
|
/Linux-v6.6/include/linux/ |
D | sbitmap.h | 413 int sbitmap_queue_init_node(struct sbitmap_queue *sbq, unsigned int depth, 421 static inline void sbitmap_queue_free(struct sbitmap_queue *sbq) in sbitmap_queue_free() argument 423 kfree(sbq->ws); in sbitmap_queue_free() 424 sbitmap_free(&sbq->sb); in sbitmap_queue_free() 435 void sbitmap_queue_recalculate_wake_batch(struct sbitmap_queue *sbq, 447 void sbitmap_queue_resize(struct sbitmap_queue *sbq, unsigned int depth); 456 int __sbitmap_queue_get(struct sbitmap_queue *sbq); 468 unsigned long __sbitmap_queue_get_batch(struct sbitmap_queue *sbq, int nr_tags, 484 int sbitmap_queue_get_shallow(struct sbitmap_queue *sbq, 496 static inline int sbitmap_queue_get(struct sbitmap_queue *sbq, in sbitmap_queue_get() argument [all …]
|
/Linux-v6.6/drivers/dma/idxd/ |
D | submit.c | 33 struct sbitmap_queue *sbq; in idxd_alloc_desc() local 38 sbq = &wq->sbq; in idxd_alloc_desc() 39 idx = sbitmap_queue_get(sbq, &cpu); in idxd_alloc_desc() 47 ws = &sbq->ws[0]; in idxd_alloc_desc() 49 sbitmap_prepare_to_wait(sbq, ws, &wait, TASK_INTERRUPTIBLE); in idxd_alloc_desc() 52 idx = sbitmap_queue_get(sbq, &cpu); in idxd_alloc_desc() 58 sbitmap_finish_wait(sbq, ws, &wait); in idxd_alloc_desc() 70 sbitmap_queue_clear(&wq->sbq, desc->id, cpu); in idxd_free_desc()
|
D | idxd.h | 221 struct sbitmap_queue sbq; member
|
D | device.c | 135 rc = sbitmap_queue_init_node(&wq->sbq, num_descs, -1, false, GFP_KERNEL, in idxd_wq_alloc_resources() 175 sbitmap_queue_free(&wq->sbq); in idxd_wq_free_resources()
|
/Linux-v6.6/drivers/net/ethernet/fungible/funcore/ |
D | fun_dev.c | 339 struct sbitmap_queue *sbq = &fdev->admin_sbq; in fun_wait_for_tag() local 340 struct sbq_wait_state *ws = &sbq->ws[0]; in fun_wait_for_tag() 345 sbitmap_prepare_to_wait(sbq, ws, &wait, TASK_UNINTERRUPTIBLE); in fun_wait_for_tag() 350 tag = sbitmap_queue_get(sbq, cpup); in fun_wait_for_tag() 356 sbitmap_finish_wait(sbq, ws, &wait); in fun_wait_for_tag()
|
/Linux-v6.6/drivers/staging/qlge/ |
D | qlge_main.c | 1145 sbq_fail = !!qlge_refill_bq(&rx_ring->sbq, gfp); in qlge_update_buffer_queues() 1154 if ((sbq_fail && QLGE_BQ_HW_OWNED(&rx_ring->sbq) < 2) || in qlge_update_buffer_queues() 1554 struct qlge_bq_desc *sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_process_mac_rx_skb() 1689 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 1719 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 1729 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 1802 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 2742 struct qlge_bq_desc *sbq_desc = &rx_ring->sbq.queue[i]; in qlge_free_sbq_buffers() 2771 if (rx_ring->sbq.queue) in qlge_free_rx_buffers() 2818 if (rx_ring->sbq.base) { in qlge_free_rx_resources() [all …]
|
D | qlge.h | 1441 offsetof(struct rx_ring, sbq) : \ 1481 struct qlge_bq sbq; member
|
/Linux-v6.6/drivers/net/ethernet/intel/ice/ |
D | ice_controlq.c | 62 struct ice_ctl_q_info *cq = &hw->sbq; in ice_sb_init_regs() 622 cq = &hw->sbq; in ice_init_ctrlq() 679 return &hw->sbq; in ice_get_sbq() 701 cq = &hw->sbq; in ice_shutdown_ctrlq() 813 ice_init_ctrlq_locks(&hw->sbq); in ice_create_all_ctrlq() 847 ice_destroy_ctrlq_locks(&hw->sbq); in ice_destroy_all_ctrlq()
|
D | ice_type.h | 870 struct ice_ctl_q_info sbq; member
|
D | ice_main.c | 1436 cq = &hw->sbq; in __ice_clean_ctrlq() 1639 if (ice_ctrlq_pending(hw, &hw->sbq)) in ice_clean_sbq_subtask() 2409 hw->sbq.num_rq_entries = ICE_SBQ_LEN; in ice_set_ctrlq_len() 2410 hw->sbq.num_sq_entries = ICE_SBQ_LEN; in ice_set_ctrlq_len() 2411 hw->sbq.rq_buf_size = ICE_SBQ_MAX_BUF_LEN; in ice_set_ctrlq_len() 2412 hw->sbq.sq_buf_size = ICE_SBQ_MAX_BUF_LEN; in ice_set_ctrlq_len()
|
/Linux-v6.6/drivers/target/iscsi/ |
D | iscsi_target_util.c | 130 struct sbitmap_queue *sbq; in iscsit_wait_for_tag() local 135 sbq = &se_sess->sess_tag_pool; in iscsit_wait_for_tag() 136 ws = &sbq->ws[0]; in iscsit_wait_for_tag() 138 sbitmap_prepare_to_wait(sbq, ws, &wait, state); in iscsit_wait_for_tag() 141 tag = sbitmap_queue_get(sbq, cpup); in iscsit_wait_for_tag() 147 sbitmap_finish_wait(sbq, ws, &wait); in iscsit_wait_for_tag()
|
/Linux-v6.6/block/ |
D | blk-mq.c | 1798 struct sbitmap_queue *sbq; in blk_mq_dispatch_wake() local 1801 sbq = &hctx->tags->bitmap_tags; in blk_mq_dispatch_wake() 1802 atomic_dec(&sbq->ws_active); in blk_mq_dispatch_wake() 1819 struct sbitmap_queue *sbq; in blk_mq_mark_tag_wait() local 1844 sbq = &hctx->tags->breserved_tags; in blk_mq_mark_tag_wait() 1846 sbq = &hctx->tags->bitmap_tags; in blk_mq_mark_tag_wait() 1847 wq = &bt_wait_ptr(sbq, hctx)->wait; in blk_mq_mark_tag_wait() 1857 atomic_inc(&sbq->ws_active); in blk_mq_mark_tag_wait() 1878 atomic_dec(&sbq->ws_active); in blk_mq_mark_tag_wait()
|
D | kyber-iosched.c | 496 khd->domain_wait[i].sbq = NULL; in kyber_init_hctx()
|