/Linux-v5.15/drivers/net/ethernet/microsoft/mana/ |
D | hw_channel.c | 326 static int mana_hwc_create_cq(struct hw_channel_context *hwc, u16 q_depth, in mana_hwc_create_cq() argument 339 eq_size = roundup_pow_of_two(GDMA_EQE_SIZE * q_depth); in mana_hwc_create_cq() 343 cq_size = roundup_pow_of_two(GDMA_CQE_SIZE * q_depth); in mana_hwc_create_cq() 366 comp_buf = kcalloc(q_depth, sizeof(struct gdma_comp), GFP_KERNEL); in mana_hwc_create_cq() 374 hwc_cq->queue_depth = q_depth; in mana_hwc_create_cq() 387 static int mana_hwc_alloc_dma_buf(struct hw_channel_context *hwc, u16 q_depth, in mana_hwc_alloc_dma_buf() argument 401 dma_buf = kzalloc(struct_size(dma_buf, reqs, q_depth), GFP_KERNEL); in mana_hwc_alloc_dma_buf() 405 dma_buf->num_reqs = q_depth; in mana_hwc_alloc_dma_buf() 407 buf_size = PAGE_ALIGN(q_depth * max_msg_size); in mana_hwc_alloc_dma_buf() 419 for (i = 0; i < q_depth; i++) { in mana_hwc_alloc_dma_buf() [all …]
|
/Linux-v5.15/drivers/net/ethernet/huawei/hinic/ |
D | hinic_hw_wq.c | 34 #define WQ_SIZE(wq) ((wq)->q_depth * (wq)->wqebb_size) 507 u16 wqebb_size, u32 wq_page_size, u16 q_depth, in hinic_wq_allocate() argument 526 if (q_depth & (q_depth - 1)) { in hinic_wq_allocate() 550 wq->q_depth = q_depth; in hinic_wq_allocate() 567 atomic_set(&wq->delta, q_depth); in hinic_wq_allocate() 568 wq->mask = q_depth - 1; in hinic_wq_allocate() 605 u16 q_depth, u16 max_wqe_size) in hinic_wqs_cmdq_alloc() argument 623 if (q_depth & (q_depth - 1)) { in hinic_wqs_cmdq_alloc() 653 wq[i].q_depth = q_depth; in hinic_wqs_cmdq_alloc() 671 atomic_set(&wq[i].delta, q_depth); in hinic_wqs_cmdq_alloc() [all …]
|
D | hinic_hw_wq.h | 30 u16 q_depth; member 80 u16 q_depth, u16 max_wqe_size); 91 u16 wqebb_size, u32 wq_page_size, u16 q_depth,
|
D | hinic_hw_cmdq.c | 368 if (next_prod_idx >= wq->q_depth) { in cmdq_sync_cmd_direct_resp() 370 next_prod_idx -= wq->q_depth; in cmdq_sync_cmd_direct_resp() 447 if (next_prod_idx >= wq->q_depth) { in cmdq_set_arm_bit() 449 next_prod_idx -= wq->q_depth; in cmdq_set_arm_bit() 754 cmdq->done = vzalloc(array_size(sizeof(*cmdq->done), wq->q_depth)); in init_cmdq() 759 wq->q_depth)); in init_cmdq()
|
D | hinic_hw_qp.c | 224 skb_arr_size = wq->q_depth * sizeof(*sq->saved_skb); in alloc_sq_skb_arr() 252 skb_arr_size = wq->q_depth * sizeof(*rq->saved_skb); in alloc_rq_skb_arr() 324 cqe_size = wq->q_depth * sizeof(*rq->cqe); in alloc_rq_cqe() 329 cqe_dma_size = wq->q_depth * sizeof(*rq->cqe_dma); in alloc_rq_cqe() 334 for (i = 0; i < wq->q_depth; i++) { in alloc_rq_cqe() 367 for (i = 0; i < wq->q_depth; i++) in free_rq_cqe()
|
/Linux-v5.15/drivers/net/ethernet/amazon/ena/ |
D | ena_eth_com.h | 80 return io_sq->q_depth - 1 - cnt; in ena_com_free_q_entries() 179 need_update = unreported_comp > (io_cq->q_depth / ENA_COMP_HEAD_THRESH); in ena_com_update_dev_comp_head() 217 if (unlikely((io_cq->head & (io_cq->q_depth - 1)) == 0)) in ena_com_cq_inc_head() 228 masked_head = io_cq->head & (io_cq->q_depth - 1); in ena_com_tx_comp_req_id_get() 246 if (unlikely(*req_id >= io_cq->q_depth)) { in ena_com_tx_comp_req_id_get()
|
D | ena_eth_com.c | 15 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_get_next_rx_cdesc() 41 tail_masked = io_sq->tail & (io_sq->q_depth - 1); in get_sq_desc_regular_queue() 56 dst_tail_mask = io_sq->tail & (io_sq->q_depth - 1); in ena_com_write_bounce_buffer_to_dev() 84 if (unlikely((io_sq->tail & (io_sq->q_depth - 1)) == 0)) in ena_com_write_bounce_buffer_to_dev() 221 if (unlikely((io_sq->tail & (io_sq->q_depth - 1)) == 0)) in ena_com_sq_update_tail() 230 idx &= (io_cq->q_depth - 1); in ena_com_rx_cdesc_idx_to_ptr() 259 head_masked = io_cq->head & (io_cq->q_depth - 1); in ena_com_cdesc_rx_pkt_get() 548 u16 q_depth = io_cq->q_depth; in ena_com_rx_pkt() local 578 if (unlikely(ena_buf[i].req_id >= q_depth)) in ena_com_rx_pkt()
|
D | ena_com.c | 89 u16 size = ADMIN_SQ_SIZE(admin_queue->q_depth); in ena_com_admin_init_sq() 112 u16 size = ADMIN_CQ_SIZE(admin_queue->q_depth); in ena_com_admin_init_cq() 135 ena_dev->aenq.q_depth = ENA_ASYNC_QUEUE_DEPTH; in ena_com_admin_init_aenq() 145 aenq->head = aenq->q_depth; in ena_com_admin_init_aenq() 155 aenq_caps |= ena_dev->aenq.q_depth & ENA_REGS_AENQ_CAPS_AENQ_DEPTH_MASK; in ena_com_admin_init_aenq() 182 if (unlikely(command_id >= admin_queue->q_depth)) { in get_comp_ctxt() 185 command_id, admin_queue->q_depth); in get_comp_ctxt() 220 queue_size_mask = admin_queue->q_depth - 1; in __ena_com_submit_admin_cmd() 226 if (cnt >= admin_queue->q_depth) { in __ena_com_submit_admin_cmd() 271 size_t size = admin_queue->q_depth * sizeof(struct ena_comp_ctx); in ena_com_init_comp_ctxt() [all …]
|
D | ena_com.h | 130 u16 q_depth; member 174 u16 q_depth; member 223 u16 q_depth; member 253 u16 q_depth; member
|
/Linux-v5.15/drivers/net/ethernet/brocade/bna/ |
D | bnad.c | 78 for (i = 0; i < ccb->q_depth; i++) { in bnad_cq_cleanup() 91 u32 q_depth, u32 index) in bnad_tx_buff_unmap() argument 114 BNA_QE_INDX_INC(index, q_depth); in bnad_tx_buff_unmap() 126 BNA_QE_INDX_INC(index, q_depth); in bnad_tx_buff_unmap() 143 for (i = 0; i < tcb->q_depth; i++) { in bnad_txq_cleanup() 147 bnad_tx_buff_unmap(bnad, unmap_q, tcb->q_depth, i); in bnad_txq_cleanup() 162 u32 wis, unmap_wis, hw_cons, cons, q_depth; in bnad_txcmpl_process() local 174 q_depth = tcb->q_depth; in bnad_txcmpl_process() 176 wis = BNA_Q_INDEX_CHANGE(cons, hw_cons, q_depth); in bnad_txcmpl_process() 177 BUG_ON(!(wis <= BNA_QE_IN_USE_CNT(tcb, tcb->q_depth))); in bnad_txcmpl_process() [all …]
|
D | bna_types.h | 422 u32 q_depth; member 550 u32 q_depth; member 566 int q_depth; member 614 u32 q_depth; member
|
D | bfa_msgq.c | 518 msgq_cfg->cmdq.q_depth = htons(msgq->cmdq.depth); in bfa_msgq_init() 520 msgq_cfg->rspq.q_depth = htons(msgq->rspq.depth); in bfa_msgq_init()
|
D | bfi.h | 413 u16 q_depth; /* Total num of entries in the queue */ member
|
D | bna_tx_rx.c | 2385 q0->rcb->q_depth = rx_cfg->q0_depth; in bna_rx_create() 2386 q0->q_depth = rx_cfg->q0_depth; in bna_rx_create() 2412 q1->rcb->q_depth = rx_cfg->q1_depth; in bna_rx_create() 2413 q1->q_depth = rx_cfg->q1_depth; in bna_rx_create() 2444 rxp->cq.ccb->q_depth = cq_depth; in bna_rx_create() 3466 txq->tcb->q_depth = tx_cfg->txq_depth; in bna_tx_create()
|
/Linux-v5.15/drivers/nvme/host/ |
D | pci.c | 33 #define SQ_SIZE(q) ((q)->q_depth << (q)->sqes) 34 #define CQ_SIZE(q) ((q)->q_depth * sizeof(struct nvme_completion)) 126 u32 q_depth; member 198 u32 q_depth; member 483 if (next_tail == nvmeq->q_depth) in nvme_write_sq_db() 507 if (++nvmeq->sq_tail == nvmeq->q_depth) in nvme_submit_cmd() 1033 if (tmp == nvmeq->q_depth) { in nvme_update_cq_head() 1146 c.create_cq.qsize = cpu_to_le16(nvmeq->q_depth - 1); in adapter_alloc_cq() 1175 c.create_sq.qsize = cpu_to_le16(nvmeq->q_depth - 1); in adapter_alloc_sq() 1444 int q_depth = dev->q_depth; in nvme_cmb_qdepth() local [all …]
|
/Linux-v5.15/drivers/block/rsxx/ |
D | cregs.c | 133 card->creg_ctrl.q_depth--; in creg_kick_queue() 185 card->creg_ctrl.q_depth++; in creg_queue_cmd() 322 card->creg_ctrl.q_depth--; in creg_reset() 399 card->creg_ctrl.q_depth + 20000); in __issue_creg_rw() 706 card->creg_ctrl.q_depth++; in rsxx_eeh_save_issued_creg()
|
D | rsxx_priv.h | 127 unsigned int q_depth; member
|
D | dma.c | 190 u32 q_depth = 0; in dma_intr_coal_auto_tune() local 198 q_depth += atomic_read(&card->ctrl[i].stats.hw_q_depth); in dma_intr_coal_auto_tune() 201 q_depth / 2, in dma_intr_coal_auto_tune()
|
/Linux-v5.15/drivers/scsi/bfa/ |
D | bfi.h | 538 #define BFI_MSGQ_FULL(_q) (((_q->pi + 1) % _q->q_depth) == _q->ci) 540 #define BFI_MSGQ_UPDATE_CI(_q) (_q->ci = (_q->ci + 1) % _q->q_depth) 541 #define BFI_MSGQ_UPDATE_PI(_q) (_q->pi = (_q->pi + 1) % _q->q_depth) 544 #define BFI_MSGQ_FREE_CNT(_q) ((_q->ci - _q->pi - 1) & (_q->q_depth - 1)) 585 u16 q_depth; /* Total num of entries in the queue */ member
|
D | bfa_fcpim.h | 119 u16 q_depth; member
|
D | bfa_defs_svc.h | 958 u16 q_depth; /* SCSI Queue depth */ member
|
/Linux-v5.15/net/mac80211/ |
D | debugfs_sta.c | 257 u32 q_depth[IEEE80211_NUM_ACS]; in sta_aql_read() local 270 q_depth[ac] = atomic_read(&sta->airtime[ac].aql_tx_pending); in sta_aql_read() 276 q_depth[0], q_depth[1], q_depth[2], q_depth[3], in sta_aql_read()
|
/Linux-v5.15/drivers/scsi/mpi3mr/ |
D | mpi3mr_os.c | 704 int q_depth) in mpi3mr_change_queue_depth() argument 711 q_depth = 1; in mpi3mr_change_queue_depth() 712 if (q_depth > shost->can_queue) in mpi3mr_change_queue_depth() 713 q_depth = shost->can_queue; in mpi3mr_change_queue_depth() 714 else if (!q_depth) in mpi3mr_change_queue_depth() 715 q_depth = MPI3MR_DEFAULT_SDEV_QD; in mpi3mr_change_queue_depth() 716 retval = scsi_change_queue_depth(sdev, q_depth); in mpi3mr_change_queue_depth() 741 mpi3mr_change_queue_depth(sdev, tgtdev->q_depth); in mpi3mr_update_sdev() 813 tgtdev->q_depth = le16_to_cpu(dev_pg0->queue_depth); in mpi3mr_update_tgtdev() 3188 mpi3mr_change_queue_depth(sdev, tgt_dev->q_depth); in mpi3mr_slave_configure()
|
D | mpi3mr.h | 450 u16 q_depth; member
|
/Linux-v5.15/drivers/s390/crypto/ |
D | ap_bus.c | 313 int *q_depth, int *q_ml, bool *q_decfg) in ap_queue_info() argument 353 *q_depth = tapq_info.tapq_gr2.qd; in ap_queue_info()
|