| /Linux-v5.15/drivers/scsi/snic/ |
| D | vnic_cq.c | 48 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, in svnic_cq_init() argument 60 iowrite32(cq_head, &cq->ctrl->cq_head); in svnic_cq_init() 75 iowrite32(0, &cq->ctrl->cq_head); in svnic_cq_clean()
|
| D | vnic_cq.h | 33 u32 cq_head; /* 0x20 */ member 105 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail,
|
| /Linux-v5.15/drivers/net/ethernet/cisco/enic/ |
| D | vnic_cq.c | 52 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, in vnic_cq_init() argument 64 iowrite32(cq_head, &cq->ctrl->cq_head); in vnic_cq_init() 81 iowrite32(0, &cq->ctrl->cq_head); in vnic_cq_clean()
|
| D | vnic_cq.h | 35 u32 cq_head; /* 0x20 */ member 117 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail,
|
| /Linux-v5.15/drivers/scsi/fnic/ |
| D | vnic_cq.c | 53 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail, in vnic_cq_init() argument 65 iowrite32(cq_head, &cq->ctrl->cq_head); in vnic_cq_init() 80 iowrite32(0, &cq->ctrl->cq_head); in vnic_cq_clean()
|
| D | vnic_cq.h | 43 u32 cq_head; /* 0x20 */ member 115 unsigned int color_enable, unsigned int cq_head, unsigned int cq_tail,
|
| /Linux-v5.15/drivers/dma/ |
| D | hisi_dma.c | 96 u32 cq_head; member 227 chan->cq_head = 0; in hisi_dma_free_chan_resources() 433 cqe = chan->cq + chan->cq_head; in hisi_dma_irq() 436 chan->cq_head = (chan->cq_head + 1) % in hisi_dma_irq() 440 chan->cq_head); in hisi_dma_irq()
|
| /Linux-v5.15/drivers/net/ethernet/marvell/octeontx2/nic/ |
| D | otx2_txrx.c | 25 cqe_hdr = (struct nix_cqe_hdr_s *)CQE_ADDR(cq, cq->cq_head); in otx2_get_next_cqe() 29 cq->cq_head++; in otx2_get_next_cqe() 30 cq->cq_head &= (cq->cqe_cnt - 1); in otx2_get_next_cqe() 322 cqe = (struct nix_cqe_rx_s *)CQE_ADDR(cq, cq->cq_head); in otx2_rx_napi_handler() 329 cq->cq_head++; in otx2_rx_napi_handler() 330 cq->cq_head &= (cq->cqe_cnt - 1); in otx2_rx_napi_handler()
|
| D | otx2_txrx.h | 124 u32 cq_head; member
|
| /Linux-v5.15/drivers/infiniband/hw/irdma/ |
| D | uk.c | 1531 u32 cq_head; in irdma_uk_clean_cq() local 1534 cq_head = cq->cq_ring.head; in irdma_uk_clean_cq() 1538 cqe = ((struct irdma_extended_cqe *)(cq->cq_base))[cq_head].buf; in irdma_uk_clean_cq() 1540 cqe = cq->cq_base[cq_head].buf; in irdma_uk_clean_cq() 1551 cq_head = (cq_head + 1) % cq->cq_ring.size; in irdma_uk_clean_cq() 1552 if (!cq_head) in irdma_uk_clean_cq()
|
| D | verbs.h | 109 u16 cq_head; member
|
| /Linux-v5.15/drivers/nvme/host/ |
| D | pci.c | 202 u16 cq_head; member 977 struct nvme_completion *hcqe = &nvmeq->cqes[nvmeq->cq_head]; in nvme_cqe_pending() 984 u16 head = nvmeq->cq_head; in nvme_ring_cq_doorbell() 1031 u32 tmp = nvmeq->cq_head + 1; in nvme_update_cq_head() 1034 nvmeq->cq_head = 0; in nvme_update_cq_head() 1037 nvmeq->cq_head = tmp; in nvme_update_cq_head() 1052 nvme_handle_cqe(nvmeq, nvmeq->cq_head); in nvme_process_cq() 1512 nvmeq->cq_head = 0; in nvme_alloc_queue() 1547 nvmeq->cq_head = 0; in nvme_init_queue()
|
| /Linux-v5.15/drivers/crypto/hisilicon/ |
| D | qm.h | 285 u16 cq_head; member
|
| D | qm.c | 776 if (qp->qp_status.cq_head == QM_Q_DEPTH - 1) { in qm_cq_head_update() 778 qp->qp_status.cq_head = 0; in qm_cq_head_update() 780 qp->qp_status.cq_head++; in qm_cq_head_update() 795 struct qm_cqe *cqe = qp->cqe + qp->qp_status.cq_head; in qm_poll_qp() 802 cqe = qp->cqe + qp->qp_status.cq_head; in qm_poll_qp() 804 qp->qp_status.cq_head, 0); in qm_poll_qp() 810 qp->qp_status.cq_head, 1); in qm_poll_qp() 943 qp_status->cq_head = 0; in qm_init_qp_status() 3014 struct qm_cqe *cqe = qp->cqe + qp->qp_status.cq_head; in hisi_qm_is_q_updated() 3021 cqe = qp->cqe + qp->qp_status.cq_head; in hisi_qm_is_q_updated()
|
| /Linux-v5.15/drivers/net/ethernet/cavium/thunder/ |
| D | nicvf_main.c | 959 u64 cq_head; in nicvf_poll() local 972 cq_head = nicvf_queue_reg_read(nic, NIC_QSET_CQ_0_7_HEAD, in nicvf_poll() 976 cq->cq_idx, cq_head); in nicvf_poll()
|