| /Linux-v5.4/drivers/scsi/lpfc/ |
| D | lpfc_debugfs.h | 637 int cq_idx; in lpfc_debug_dump_cq_by_id() local 639 for (cq_idx = 0; cq_idx < phba->cfg_hdw_queue; cq_idx++) in lpfc_debug_dump_cq_by_id() 640 if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid) in lpfc_debug_dump_cq_by_id() 643 if (cq_idx < phba->cfg_hdw_queue) { in lpfc_debug_dump_cq_by_id() 644 pr_err("IO CQ[Idx:%d|Qid:%d]\n", cq_idx, qid); in lpfc_debug_dump_cq_by_id() 645 lpfc_debug_dump_q(phba->sli4_hba.hdwq[cq_idx].io_cq); in lpfc_debug_dump_cq_by_id()
|
| /Linux-v5.4/drivers/net/ethernet/mellanox/mlx4/ |
| D | en_cq.c | 90 int cq_idx) in mlx4_en_activate_cq() argument 128 cq_idx = cq_idx % priv->rx_ring_num; in mlx4_en_activate_cq() 129 rx_cq = priv->rx_cq[cq_idx]; in mlx4_en_activate_cq()
|
| D | mlx4_en.h | 694 int cq_idx);
|
| /Linux-v5.4/drivers/net/ethernet/amazon/ena/ |
| D | ena_admin_defs.h | 244 u16 cq_idx; member 332 u16 cq_idx; member 347 u16 cq_idx; member 820 u16 cq_idx; member
|
| D | ena_com.c | 1161 struct ena_com_io_sq *io_sq, u16 cq_idx) in ena_com_create_io_sq() argument 1192 create_cmd.cq_idx = cq_idx; in ena_com_create_io_sq() 1251 rss->rss_ind_tbl[i].cq_idx = io_sq->idx; in ena_com_ind_tbl_convert_to_device() 1268 if (rss->rss_ind_tbl[i].cq_idx > ENA_TOTAL_NUM_QUEUES) in ena_com_ind_tbl_convert_from_device() 1270 idx = (u8)rss->rss_ind_tbl[i].cq_idx; in ena_com_ind_tbl_convert_from_device() 1384 io_cq->idx = cmd_completion.cq_idx; in ena_com_create_io_cq() 1464 destroy_cmd.cq_idx = io_cq->idx; in ena_com_destroy_io_cq()
|
| /Linux-v5.4/drivers/infiniband/hw/efa/ |
| D | efa_verbs.c | 707 create_qp_params.send_cq_idx = to_ecq(init_attr->send_cq)->cq_idx; in efa_create_qp() 708 create_qp_params.recv_cq_idx = to_ecq(init_attr->recv_cq)->cq_idx; in efa_create_qp() 881 static int efa_destroy_cq_idx(struct efa_dev *dev, int cq_idx) in efa_destroy_cq_idx() argument 883 struct efa_com_destroy_cq_params params = { .cq_idx = cq_idx }; in efa_destroy_cq_idx() 895 cq->cq_idx, cq->cpu_addr, cq->size, &cq->dma_addr); in efa_destroy_cq() 897 efa_destroy_cq_idx(dev, cq->cq_idx); in efa_destroy_cq() 1004 resp.cq_idx = result.cq_idx; in efa_create_cq() 1005 cq->cq_idx = result.cq_idx; in efa_create_cq() 1012 cq->cq_idx); in efa_create_cq() 1029 cq->cq_idx, result.actual_depth, &cq->dma_addr, cq->cpu_addr); in efa_create_cq() [all …]
|
| D | efa_com_cmd.h | 79 u16 cq_idx; member 85 u16 cq_idx; member
|
| D | efa_com_cmd.c | 185 result->cq_idx = cmd_completion.cq_idx; in efa_com_create_cq() 199 destroy_cmd.cq_idx = params->cq_idx; in efa_com_destroy_cq() 211 params->cq_idx, err); in efa_com_destroy_cq()
|
| D | efa_admin_cmds_defs.h | 466 u16 cq_idx; member 475 u16 cq_idx; member
|
| D | efa.h | 95 u16 cq_idx; member
|
| /Linux-v5.4/drivers/net/ethernet/cavium/thunder/ |
| D | nicvf_main.c | 848 static int nicvf_cq_intr_handler(struct net_device *netdev, u8 cq_idx, in nicvf_cq_intr_handler() argument 856 struct cmp_queue *cq = &qs->cq[cq_idx]; in nicvf_cq_intr_handler() 859 struct snd_queue *sq = &qs->sq[cq_idx]; in nicvf_cq_intr_handler() 860 struct rcv_queue *rq = &qs->rq[cq_idx]; in nicvf_cq_intr_handler() 867 cqe_count = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_STATUS, cq_idx); in nicvf_cq_intr_handler() 873 cqe_head = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_HEAD, cq_idx) >> 9; in nicvf_cq_intr_handler() 914 cq_idx, processed_cqe); in nicvf_cq_intr_handler() 924 txq_idx = nicvf_netdev_qidx(nic, cq_idx); in nicvf_cq_intr_handler() 928 nicvf_xdp_sq_doorbell(nic, sq, cq_idx); in nicvf_cq_intr_handler() 968 work_done = nicvf_cq_intr_handler(netdev, cq->cq_idx, napi, budget); in nicvf_poll() [all …]
|
| D | nicvf_queues.h | 245 u8 cq_idx; /* CQ index (0 to 7) in the QS */ member 267 u8 cq_idx; /* CQ index (0 to 7) in the above QS */ member
|
| D | nic.h | 123 u8 cq_idx; /* Completion queue index */ member
|
| D | nicvf_queues.c | 764 rq->cq_idx = qidx; in nicvf_rcv_queue_config() 780 (rq->cq_idx << 16) | (rq->cont_rbdr_qs << 9) | in nicvf_rcv_queue_config() 877 sq->cq_idx = qidx; in nicvf_snd_queue_config() 884 mbx.sq.cfg = (sq->cq_qs << 3) | sq->cq_idx; in nicvf_snd_queue_config()
|
| /Linux-v5.4/drivers/infiniband/hw/usnic/ |
| D | usnic_abi.h | 83 u32 cq_idx[USNIC_QP_GRP_MAX_CQS]; member
|
| D | usnic_ib_verbs.c | 155 resp.cq_idx[i] = chunk->res[i]->vnic_idx; in usnic_ib_fill_create_qp_resp()
|
| /Linux-v5.4/include/uapi/rdma/ |
| D | efa-abi.h | 54 __u16 cq_idx; member
|
| /Linux-v5.4/drivers/scsi/snic/ |
| D | snic_scsi.c | 1201 unsigned int cq_idx, in snic_io_cmpl_handler() argument 1284 unsigned int cq_idx; in snic_fwcq_cmpl_handler() local 1288 for (cq_idx = snic->wq_count; cq_idx < snic->cq_count; cq_idx++) { in snic_fwcq_cmpl_handler() 1289 nent_per_cq = vnic_cq_fw_service(&snic->cq[cq_idx], in snic_fwcq_cmpl_handler()
|
| /Linux-v5.4/drivers/scsi/qedi/ |
| D | qedi_fw.c | 2053 u16 cq_idx; in qedi_iscsi_send_ioreq() local 2081 cq_idx = smp_processor_id() % qedi->num_queues; in qedi_iscsi_send_ioreq() 2146 task_params.cq_rss_number = cq_idx; in qedi_iscsi_send_ioreq()
|
| /Linux-v5.4/drivers/nvme/host/ |
| D | pci.c | 149 static inline unsigned int cq_idx(unsigned int qid, u32 stride) in cq_idx() function 278 nvmeq->dbbuf_cq_db = &dev->dbbuf_dbs[cq_idx(qid, dev->db_stride)]; in nvme_dbbuf_init() 280 nvmeq->dbbuf_cq_ei = &dev->dbbuf_eis[cq_idx(qid, dev->db_stride)]; in nvme_dbbuf_init()
|
| /Linux-v5.4/drivers/infiniband/hw/ocrdma/ |
| D | ocrdma_hw.c | 965 static void ocrdma_qp_cq_handler(struct ocrdma_dev *dev, u16 cq_idx) in ocrdma_qp_cq_handler() argument 970 if (cq_idx >= OCRDMA_MAX_CQ) in ocrdma_qp_cq_handler() 973 cq = dev->cq_tbl[cq_idx]; in ocrdma_qp_cq_handler()
|
| /Linux-v5.4/drivers/scsi/qedf/ |
| D | qedf_io.c | 605 uint16_t cq_idx = smp_processor_id() % qedf->num_queues; in qedf_init_task() local 638 io_req->task_params->cq_rss_number = cq_idx; in qedf_init_task()
|
| /Linux-v5.4/drivers/net/ethernet/chelsio/cxgb3/ |
| D | t3_cpl.h | 234 __be16 cq_idx; member
|