/Linux-v6.1/lib/ |
D | sbitmap.c | 391 static unsigned int sbq_calc_wake_batch(struct sbitmap_queue *sbq, in sbq_calc_wake_batch() argument 413 shallow_depth = min(1U << sbq->sb.shift, sbq->min_shallow_depth); in sbq_calc_wake_batch() 414 depth = ((depth >> sbq->sb.shift) * shallow_depth + in sbq_calc_wake_batch() 415 min(depth & ((1U << sbq->sb.shift) - 1), shallow_depth)); in sbq_calc_wake_batch() 422 int sbitmap_queue_init_node(struct sbitmap_queue *sbq, unsigned int depth, in sbitmap_queue_init_node() argument 428 ret = sbitmap_init_node(&sbq->sb, depth, shift, flags, node, in sbitmap_queue_init_node() 433 sbq->min_shallow_depth = UINT_MAX; in sbitmap_queue_init_node() 434 sbq->wake_batch = sbq_calc_wake_batch(sbq, depth); in sbitmap_queue_init_node() 435 atomic_set(&sbq->wake_index, 0); in sbitmap_queue_init_node() 436 atomic_set(&sbq->ws_active, 0); in sbitmap_queue_init_node() [all …]
|
/Linux-v6.1/include/linux/ |
D | sbitmap.h | 407 int sbitmap_queue_init_node(struct sbitmap_queue *sbq, unsigned int depth, 415 static inline void sbitmap_queue_free(struct sbitmap_queue *sbq) in sbitmap_queue_free() argument 417 kfree(sbq->ws); in sbitmap_queue_free() 418 sbitmap_free(&sbq->sb); in sbitmap_queue_free() 429 void sbitmap_queue_recalculate_wake_batch(struct sbitmap_queue *sbq, 441 void sbitmap_queue_resize(struct sbitmap_queue *sbq, unsigned int depth); 450 int __sbitmap_queue_get(struct sbitmap_queue *sbq); 462 unsigned long __sbitmap_queue_get_batch(struct sbitmap_queue *sbq, int nr_tags, 478 int sbitmap_queue_get_shallow(struct sbitmap_queue *sbq, 490 static inline int sbitmap_queue_get(struct sbitmap_queue *sbq, in sbitmap_queue_get() argument [all …]
|
/Linux-v6.1/drivers/dma/idxd/ |
D | submit.c | 33 struct sbitmap_queue *sbq; in idxd_alloc_desc() local 38 sbq = &wq->sbq; in idxd_alloc_desc() 39 idx = sbitmap_queue_get(sbq, &cpu); in idxd_alloc_desc() 47 ws = &sbq->ws[0]; in idxd_alloc_desc() 49 sbitmap_prepare_to_wait(sbq, ws, &wait, TASK_INTERRUPTIBLE); in idxd_alloc_desc() 52 idx = sbitmap_queue_get(sbq, &cpu); in idxd_alloc_desc() 58 sbitmap_finish_wait(sbq, ws, &wait); in idxd_alloc_desc() 70 sbitmap_queue_clear(&wq->sbq, desc->id, cpu); in idxd_free_desc()
|
D | idxd.h | 212 struct sbitmap_queue sbq; member
|
D | device.c | 136 rc = sbitmap_queue_init_node(&wq->sbq, num_descs, -1, false, GFP_KERNEL, in idxd_wq_alloc_resources() 176 sbitmap_queue_free(&wq->sbq); in idxd_wq_free_resources()
|
/Linux-v6.1/drivers/net/ethernet/fungible/funcore/ |
D | fun_dev.c | 340 struct sbitmap_queue *sbq = &fdev->admin_sbq; in fun_wait_for_tag() local 341 struct sbq_wait_state *ws = &sbq->ws[0]; in fun_wait_for_tag() 346 sbitmap_prepare_to_wait(sbq, ws, &wait, TASK_UNINTERRUPTIBLE); in fun_wait_for_tag() 351 tag = sbitmap_queue_get(sbq, cpup); in fun_wait_for_tag() 357 sbitmap_finish_wait(sbq, ws, &wait); in fun_wait_for_tag()
|
/Linux-v6.1/drivers/staging/qlge/ |
D | qlge_main.c | 1145 sbq_fail = !!qlge_refill_bq(&rx_ring->sbq, gfp); in qlge_update_buffer_queues() 1154 if ((sbq_fail && QLGE_BQ_HW_OWNED(&rx_ring->sbq) < 2) || in qlge_update_buffer_queues() 1554 struct qlge_bq_desc *sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_process_mac_rx_skb() 1689 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 1719 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 1729 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 1802 sbq_desc = qlge_get_curr_buf(&rx_ring->sbq); in qlge_build_rx_skb() 2742 struct qlge_bq_desc *sbq_desc = &rx_ring->sbq.queue[i]; in qlge_free_sbq_buffers() 2771 if (rx_ring->sbq.queue) in qlge_free_rx_buffers() 2818 if (rx_ring->sbq.base) { in qlge_free_rx_resources() [all …]
|
D | qlge.h | 1441 offsetof(struct rx_ring, sbq) : \ 1481 struct qlge_bq sbq; member
|
/Linux-v6.1/drivers/net/ethernet/intel/ice/ |
D | ice_controlq.c | 62 struct ice_ctl_q_info *cq = &hw->sbq; in ice_sb_init_regs() 623 cq = &hw->sbq; in ice_init_ctrlq() 683 return &hw->sbq; in ice_get_sbq() 705 cq = &hw->sbq; in ice_shutdown_ctrlq() 817 ice_init_ctrlq_locks(&hw->sbq); in ice_create_all_ctrlq() 851 ice_destroy_ctrlq_locks(&hw->sbq); in ice_destroy_all_ctrlq()
|
D | ice_type.h | 858 struct ice_ctl_q_info sbq; member
|
D | ice_main.c | 1405 cq = &hw->sbq; in __ice_clean_ctrlq() 1603 if (ice_ctrlq_pending(hw, &hw->sbq)) in ice_clean_sbq_subtask() 2376 hw->sbq.num_rq_entries = ICE_SBQ_LEN; in ice_set_ctrlq_len() 2377 hw->sbq.num_sq_entries = ICE_SBQ_LEN; in ice_set_ctrlq_len() 2378 hw->sbq.rq_buf_size = ICE_SBQ_MAX_BUF_LEN; in ice_set_ctrlq_len() 2379 hw->sbq.sq_buf_size = ICE_SBQ_MAX_BUF_LEN; in ice_set_ctrlq_len()
|
/Linux-v6.1/drivers/target/iscsi/ |
D | iscsi_target_util.c | 130 struct sbitmap_queue *sbq; in iscsit_wait_for_tag() local 135 sbq = &se_sess->sess_tag_pool; in iscsit_wait_for_tag() 136 ws = &sbq->ws[0]; in iscsit_wait_for_tag() 138 sbitmap_prepare_to_wait(sbq, ws, &wait, state); in iscsit_wait_for_tag() 141 tag = sbitmap_queue_get(sbq, cpup); in iscsit_wait_for_tag() 147 sbitmap_finish_wait(sbq, ws, &wait); in iscsit_wait_for_tag()
|
/Linux-v6.1/block/ |
D | blk-mq.c | 1743 struct sbitmap_queue *sbq; in blk_mq_dispatch_wake() local 1746 sbq = &hctx->tags->bitmap_tags; in blk_mq_dispatch_wake() 1747 atomic_dec(&sbq->ws_active); in blk_mq_dispatch_wake() 1764 struct sbitmap_queue *sbq = &hctx->tags->bitmap_tags; in blk_mq_mark_tag_wait() local 1787 wq = &bt_wait_ptr(sbq, hctx)->wait; in blk_mq_mark_tag_wait() 1797 atomic_inc(&sbq->ws_active); in blk_mq_mark_tag_wait() 1818 atomic_dec(&sbq->ws_active); in blk_mq_mark_tag_wait()
|
D | kyber-iosched.c | 498 khd->domain_wait[i].sbq = NULL; in kyber_init_hctx()
|