/Linux-v4.19/drivers/net/ethernet/cavium/thunder/ |
D | nicvf_queues.c | 34 static int nicvf_poll_reg(struct nicvf *nic, int qidx, in nicvf_poll_reg() argument 45 reg_val = nicvf_queue_reg_read(nic, reg, qidx); in nicvf_poll_reg() 509 struct snd_queue *sq, int q_len, int qidx) in nicvf_init_snd_queue() argument 529 qidx += ((nic->sqs_id + 1) * MAX_SND_QUEUES_PER_QS); in nicvf_init_snd_queue() 530 if (qidx < nic->pnicvf->xdp_tx_queues) { in nicvf_init_snd_queue() 630 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() argument 633 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, 0); in nicvf_reclaim_snd_queue() 635 if (nicvf_poll_reg(nic, qidx, NIC_QSET_SQ_0_7_STATUS, 21, 1, 0x01)) in nicvf_reclaim_snd_queue() 638 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, NICVF_SQ_RESET); in nicvf_reclaim_snd_queue() 642 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() argument [all …]
|
D | nicvf_main.c | 74 static inline u8 nicvf_netdev_qidx(struct nicvf *nic, u8 qidx) in nicvf_netdev_qidx() argument 77 return qidx + ((nic->sqs_id + 1) * MAX_CMP_QUEUES_PER_QS); in nicvf_netdev_qidx() 79 return qidx; in nicvf_netdev_qidx() 103 u64 qidx, u64 val) in nicvf_queue_reg_write() argument 107 writeq_relaxed(val, addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_write() 110 u64 nicvf_queue_reg_read(struct nicvf *nic, u64 offset, u64 qidx) in nicvf_queue_reg_read() argument 114 return readq_relaxed(addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_read() 971 int qidx; in nicvf_handle_qs_err() local 977 for (qidx = 0; qidx < qs->cq_cnt; qidx++) { in nicvf_handle_qs_err() 979 qidx); in nicvf_handle_qs_err() [all …]
|
D | nicvf_ethtool.c | 218 int stats, qidx; in nicvf_get_qset_strings() local 221 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_strings() 223 sprintf(*data, "rxq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings() 229 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_strings() 231 sprintf(*data, "txq%d: %s", qidx + start_qidx, in nicvf_get_qset_strings() 305 int stat, qidx; in nicvf_get_qset_stats() local 310 for (qidx = 0; qidx < nic->qs->rq_cnt; qidx++) { in nicvf_get_qset_stats() 311 nicvf_update_rq_stats(nic, qidx); in nicvf_get_qset_stats() 313 *((*data)++) = ((u64 *)&nic->qs->rq[qidx].stats) in nicvf_get_qset_stats() 317 for (qidx = 0; qidx < nic->qs->sq_cnt; qidx++) { in nicvf_get_qset_stats() [all …]
|
D | nicvf_queues.h | 339 int qidx, bool enable); 341 void nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx); 342 void nicvf_sq_disable(struct nicvf *nic, int qidx); 345 struct snd_queue *sq, int qidx); 368 u64 qidx, u64 val); 370 u64 offset, u64 qidx);
|
/Linux-v4.19/drivers/crypto/chelsio/ |
D | chcr_ipsec.c | 353 u32 ctrl0, qidx; in copy_cpltx_pktxt() local 359 qidx = skb->queue_mapping; in copy_cpltx_pktxt() 360 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_cpltx_pktxt() 395 unsigned int qidx; in copy_key_cpltx_pktxt() local 399 qidx = skb->queue_mapping; in copy_key_cpltx_pktxt() 400 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_key_cpltx_pktxt() 446 int qidx = skb_get_queue_mapping(skb); in chcr_crypto_wreq() local 447 struct sge_eth_txq *q = &adap->sge.ethtxq[qidx + pi->first_qset]; in chcr_crypto_wreq() 567 int qidx, left, credits; in chcr_ipsec_xmit() local 588 qidx = skb->queue_mapping; in chcr_ipsec_xmit() [all …]
|
/Linux-v4.19/drivers/net/ethernet/broadcom/bnxt/ |
D | bnxt_dcb.c | 51 u8 qidx; in bnxt_hwrm_queue_pri2cos_cfg() local 56 qidx = bp->tc_to_qidx[ets->prio_tc[i]]; in bnxt_hwrm_queue_pri2cos_cfg() 57 pri2cos[i] = bp->q_info[qidx].queue_id; in bnxt_hwrm_queue_pri2cos_cfg() 101 u8 qidx = bp->tc_to_qidx[i]; in bnxt_hwrm_queue_cos2bw_cfg() local 105 qidx); in bnxt_hwrm_queue_cos2bw_cfg() 108 cos2bw.queue_id = bp->q_info[qidx].queue_id; in bnxt_hwrm_queue_cos2bw_cfg() 124 data = &req.unused_0 + qidx * (sizeof(cos2bw) - 4); in bnxt_hwrm_queue_cos2bw_cfg() 126 if (qidx == 0) { in bnxt_hwrm_queue_cos2bw_cfg() 260 u8 qidx = bp->tc_to_qidx[i]; in bnxt_hwrm_queue_pfc_cfg() local 262 if (!BNXT_LLQ(bp->q_info[qidx].queue_profile)) { in bnxt_hwrm_queue_pfc_cfg()
|
/Linux-v4.19/drivers/scsi/qla2xxx/ |
D | qla_nvme.c | 85 unsigned int qidx, u16 qsize, void **handle) in qla_nvme_alloc_queue() argument 91 if (!qidx) in qla_nvme_alloc_queue() 92 qidx++; in qla_nvme_alloc_queue() 99 __func__, handle, qidx, qsize); in qla_nvme_alloc_queue() 101 if (qidx > qla_nvme_fc_transport.max_hw_queues) { in qla_nvme_alloc_queue() 104 __func__, qidx, qla_nvme_fc_transport.max_hw_queues); in qla_nvme_alloc_queue() 108 if (ha->queue_pair_map[qidx]) { in qla_nvme_alloc_queue() 109 *handle = ha->queue_pair_map[qidx]; in qla_nvme_alloc_queue() 112 *handle, qidx); in qla_nvme_alloc_queue()
|
/Linux-v4.19/drivers/scsi/lpfc/ |
D | lpfc_nvme.h | 39 uint32_t qidx; /* queue index passed to create */ member 93 uint16_t qidx; member
|
D | lpfc_debugfs.c | 3181 int qidx; in lpfc_idiag_wqs_for_cq() local 3183 for (qidx = 0; qidx < phba->cfg_fcp_io_channel; qidx++) { in lpfc_idiag_wqs_for_cq() 3184 qp = phba->sli4_hba.fcp_wq[qidx]; in lpfc_idiag_wqs_for_cq() 3191 for (qidx = 0; qidx < phba->cfg_nvme_io_channel; qidx++) { in lpfc_idiag_wqs_for_cq() 3192 qp = phba->sli4_hba.nvme_wq[qidx]; in lpfc_idiag_wqs_for_cq() 3261 int qidx, rc; in lpfc_idiag_cqs_for_eq() local 3263 for (qidx = 0; qidx < phba->cfg_fcp_io_channel; qidx++) { in lpfc_idiag_cqs_for_eq() 3264 qp = phba->sli4_hba.fcp_cq[qidx]; in lpfc_idiag_cqs_for_eq() 3282 for (qidx = 0; qidx < phba->cfg_nvme_io_channel; qidx++) { in lpfc_idiag_cqs_for_eq() 3283 qp = phba->sli4_hba.nvme_cq[qidx]; in lpfc_idiag_cqs_for_eq() [all …]
|
D | lpfc_init.c | 8688 int qidx, uint32_t qtype) in lpfc_create_wq_cq() argument 8696 ((eq) ? ((cq) ? "WQ" : "CQ") : "EQ"), qidx); in lpfc_create_wq_cq() 8706 qidx, (uint32_t)rc); in lpfc_create_wq_cq() 8709 cq->chann = qidx; in lpfc_create_wq_cq() 8718 qidx, cq->queue_id, qidx, eq->queue_id); in lpfc_create_wq_cq() 8725 qidx, (uint32_t)rc); in lpfc_create_wq_cq() 8729 wq->chann = qidx; in lpfc_create_wq_cq() 8738 qidx, wq->queue_id, wq->assoc_qid, qidx, cq->queue_id); in lpfc_create_wq_cq() 8776 int qidx; in lpfc_sli4_queue_setup() local 8837 for (qidx = 0; qidx < io_channel; qidx++) { in lpfc_sli4_queue_setup() [all …]
|
D | lpfc_debugfs.h | 515 lpfc_debug_dump_hba_eq(struct lpfc_hba *phba, int qidx) in lpfc_debug_dump_hba_eq() argument 519 qp = phba->sli4_hba.hba_eq[qidx]; in lpfc_debug_dump_hba_eq() 521 pr_err("EQ[Idx:%d|Qid:%d]\n", qidx, qp->queue_id); in lpfc_debug_dump_hba_eq()
|
D | lpfc_nvme.c | 215 unsigned int qidx, u16 qsize, in lpfc_nvme_create_queue() argument 233 qhandle->qidx = qidx; in lpfc_nvme_create_queue() 239 if (qidx) { in lpfc_nvme_create_queue() 241 qhandle->index = ((qidx - 1) % in lpfc_nvme_create_queue() 245 qhandle->index = qidx; in lpfc_nvme_create_queue() 251 qidx, qhandle->cpu_id, qhandle->index, qhandle); in lpfc_nvme_create_queue() 272 unsigned int qidx, in lpfc_nvme_delete_queue() argument 286 lport, qidx, handle); in lpfc_nvme_delete_queue() 1536 if (!lpfc_queue_info->qidx && !pnvme_fcreq->sg_cnt) { in lpfc_nvme_fcp_io_submit() 1639 if (lpfc_queue_info->qidx) { in lpfc_nvme_fcp_io_submit()
|
D | lpfc_sli.c | 86 struct lpfc_eqe *eqe, uint32_t qidx); 5510 int qidx; in lpfc_sli4_arm_cqeq_intr() local 5520 for (qidx = 0; qidx < phba->cfg_fcp_io_channel; qidx++) in lpfc_sli4_arm_cqeq_intr() 5521 sli4_hba->sli4_cq_release(sli4_hba->fcp_cq[qidx], in lpfc_sli4_arm_cqeq_intr() 5525 for (qidx = 0; qidx < phba->cfg_nvme_io_channel; qidx++) in lpfc_sli4_arm_cqeq_intr() 5526 sli4_hba->sli4_cq_release(sli4_hba->nvme_cq[qidx], in lpfc_sli4_arm_cqeq_intr() 5533 for (qidx = 0; qidx < phba->io_channel_irqs; qidx++) in lpfc_sli4_arm_cqeq_intr() 5534 sli4_hba->sli4_eq_release(sli4_hba->hba_eq[qidx], in lpfc_sli4_arm_cqeq_intr() 5538 for (qidx = 0; qidx < phba->cfg_nvmet_mrq; qidx++) { in lpfc_sli4_arm_cqeq_intr() 5540 sli4_hba->nvmet_cqset[qidx], in lpfc_sli4_arm_cqeq_intr() [all …]
|
/Linux-v4.19/drivers/scsi/csiostor/ |
D | csio_wr.c | 744 csio_wr_cleanup_eq_stpg(struct csio_hw *hw, int qidx) in csio_wr_cleanup_eq_stpg() argument 746 struct csio_q *q = csio_hw_to_wrm(hw)->q_arr[qidx]; in csio_wr_cleanup_eq_stpg() 761 csio_wr_cleanup_iq_ftr(struct csio_hw *hw, int qidx) in csio_wr_cleanup_iq_ftr() argument 764 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_cleanup_iq_ftr() 860 csio_wr_get(struct csio_hw *hw, int qidx, uint32_t size, in csio_wr_get() argument 864 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_get() 875 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_get() 979 csio_wr_issue(struct csio_hw *hw, int qidx, bool prio) in csio_wr_issue() argument 982 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_issue() 984 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_issue() [all …]
|
/Linux-v4.19/include/linux/ |
D | nvme-fc-driver.h | 402 unsigned int qidx, u16 qsize, 405 unsigned int qidx, void *handle);
|
/Linux-v4.19/drivers/net/ethernet/chelsio/cxgb3/ |
D | cxgb3_main.c | 404 int i, j, err, qidx = 0; in request_msix_data_irqs() local 410 err = request_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs() 412 adap->sge.qs[qidx]. in request_msix_data_irqs() 414 adap->msix_info[qidx + 1].desc, in request_msix_data_irqs() 415 &adap->sge.qs[qidx]); in request_msix_data_irqs() 417 while (--qidx >= 0) in request_msix_data_irqs() 418 free_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs() 419 &adap->sge.qs[qidx]); in request_msix_data_irqs() 422 qidx++; in request_msix_data_irqs() 943 static int send_pktsched_cmd(struct adapter *adap, int sched, int qidx, int lo, in send_pktsched_cmd() argument [all …]
|
D | sge.c | 167 static inline struct sge_qset *fl_to_qset(const struct sge_fl *q, int qidx) in fl_to_qset() argument 169 return container_of(q, struct sge_qset, fl[qidx]); in fl_to_qset() 177 static inline struct sge_qset *txq_to_qset(const struct sge_txq *q, int qidx) in txq_to_qset() argument 179 return container_of(q, struct sge_qset, txq[qidx]); in txq_to_qset() 1265 int qidx; in t3_eth_xmit() local 1283 qidx = skb_get_queue_mapping(skb); in t3_eth_xmit() 1284 qs = &pi->qs[qidx]; in t3_eth_xmit() 1286 txq = netdev_get_tx_queue(dev, qidx); in t3_eth_xmit()
|
/Linux-v4.19/drivers/crypto/ccp/ |
D | ccp-dev-v5.c | 222 n = head_idx + COMMANDS_PER_QUEUE - cmd_q->qidx - 1; in ccp5_get_free_slots() 244 mP = (u32 *) &cmd_q->qbase[cmd_q->qidx]; in ccp5_do_cmd() 249 cmd_q->qidx = (cmd_q->qidx + 1) % COMMANDS_PER_QUEUE; in ccp5_do_cmd() 255 tail = low_address(cmd_q->qdma_tail + cmd_q->qidx * Q_DESC_SIZE); in ccp5_do_cmd() 834 cmd_q->qidx = 0; in ccp5_init()
|
D | ccp-dev.h | 277 unsigned int qidx; member
|
/Linux-v4.19/drivers/net/ethernet/broadcom/bnx2x/ |
D | bnx2x_sriov.c | 1742 int qidx = 0, abs_vfid; in bnx2x_iov_eq_sp_event() local 1791 qidx = cid & ((1 << BNX2X_VF_CID_WND)-1); in bnx2x_iov_eq_sp_event() 1805 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event() 1806 vfq_get(vf, qidx)->sp_obj.complete_cmd(bp, in bnx2x_iov_eq_sp_event() 1808 qidx)->sp_obj, in bnx2x_iov_eq_sp_event() 1813 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event() 1814 bnx2x_vf_handle_classification_eqe(bp, vfq_get(vf, qidx), elem); in bnx2x_iov_eq_sp_event() 1818 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event() 1823 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event() 1828 vf->abs_vfid, qidx); in bnx2x_iov_eq_sp_event()
|
D | bnx2x_vfpf.c | 26 static int bnx2x_vfpf_teardown_queue(struct bnx2x *bp, int qidx); 686 static int bnx2x_vfpf_teardown_queue(struct bnx2x *bp, int qidx) in bnx2x_vfpf_teardown_queue() argument 696 req->vf_qid = qidx; in bnx2x_vfpf_teardown_queue() 708 BNX2X_ERR("Sending TEARDOWN for queue %d failed: %d\n", qidx, in bnx2x_vfpf_teardown_queue() 715 BNX2X_ERR("TEARDOWN for queue %d failed: %d\n", qidx, in bnx2x_vfpf_teardown_queue()
|
/Linux-v4.19/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 1161 int qidx, credits, max_pkt_len; in t4vf_eth_xmit() local 1196 qidx = skb_get_queue_mapping(skb); in t4vf_eth_xmit() 1197 BUG_ON(qidx >= pi->nqsets); in t4vf_eth_xmit() 1198 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in t4vf_eth_xmit() 1229 dev->name, qidx); in t4vf_eth_xmit()
|
/Linux-v4.19/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
D | fwsignal.c | 1975 int qidx, hslot; in brcmf_fws_rollback_toq() local 1980 qidx = 2 * fifo; in brcmf_fws_rollback_toq() 1982 qidx++; in brcmf_fws_rollback_toq() 1984 pktout = brcmu_pktq_penq_head(&entry->psq, qidx, skb); in brcmf_fws_rollback_toq() 1986 brcmf_err("%s queue %d full\n", entry->name, qidx); in brcmf_fws_rollback_toq()
|
/Linux-v4.19/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 1301 int qidx, credits; in cxgb4_eth_xmit() local 1344 qidx = skb_get_queue_mapping(skb); in cxgb4_eth_xmit() 1356 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in cxgb4_eth_xmit() 1381 dev->name, qidx); in cxgb4_eth_xmit() 1636 int qidx, credits, max_pkt_len; in cxgb4_vf_eth_xmit() local 1667 qidx = skb_get_queue_mapping(skb); in cxgb4_vf_eth_xmit() 1668 WARN_ON(qidx >= pi->nqsets); in cxgb4_vf_eth_xmit() 1669 txq = &adapter->sge.ethtxq[pi->first_qset + qidx]; in cxgb4_vf_eth_xmit() 1693 dev->name, qidx); in cxgb4_vf_eth_xmit()
|
/Linux-v4.19/drivers/nvme/host/ |
D | fc.c | 1802 unsigned int qidx) in __nvme_fc_init_hctx() argument 1804 struct nvme_fc_queue *queue = &ctrl->queues[qidx]; in __nvme_fc_init_hctx() 1888 struct nvme_fc_queue *queue, unsigned int qidx) in __nvme_fc_delete_hw_queue() argument 1891 ctrl->lport->ops->delete_queue(&ctrl->lport->localport, qidx, in __nvme_fc_delete_hw_queue() 1907 struct nvme_fc_queue *queue, unsigned int qidx, u16 qsize) in __nvme_fc_create_hw_queue() argument 1914 qidx, qsize, &queue->lldd_handle); in __nvme_fc_create_hw_queue()
|