Home
last modified time | relevance | path

Searched refs:cqes (Results 1 – 10 of 10) sorted by relevance

/Linux-v4.19/drivers/infiniband/hw/cxgb4/
Drestrack.c337 struct t4_cqe *cqes) in fill_hwcqes() argument
342 if (fill_cqe(msg, cqes, idx, "hwcq_idx")) in fill_hwcqes()
345 if (fill_cqe(msg, cqes + 1, idx, "hwcq_idx")) in fill_hwcqes()
354 struct t4_cqe *cqes) in fill_swcqes() argument
362 if (fill_cqe(msg, cqes, idx, "swcq_idx")) in fill_swcqes()
367 if (fill_cqe(msg, cqes + 1, idx, "swcq_idx")) in fill_swcqes()
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/
Den_stats.h229 u64 cqes ____cacheline_aligned_in_smp;
239 u64 cqes ____cacheline_aligned_in_smp;
Den_stats.c157 s->rx_xdp_tx_cqe += xdpsq_stats->cqes; in mlx5e_grp_sw_update_stats()
180 s->tx_xdp_cqes += xdpsq_red_stats->cqes; in mlx5e_grp_sw_update_stats()
207 s->tx_cqes += sq_stats->cqes; in mlx5e_grp_sw_update_stats()
1179 { MLX5E_DECLARE_TX_STAT(struct mlx5e_sq_stats, cqes) },
1188 { MLX5E_DECLARE_RQ_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) },
1195 { MLX5E_DECLARE_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) },
Den_tx.c563 stats->cqes += i; in mlx5e_poll_tx_cq()
/Linux-v4.19/drivers/net/ethernet/broadcom/
Dcnic.c1425 struct kcqe *cqes[], u32 num_cqes) in cnic_reply_bnx2x_kcqes() argument
1434 cqes, num_cqes); in cnic_reply_bnx2x_kcqes()
1552 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_init2() local
1582 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_init2()
1583 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_init2()
1881 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_ofld1() local
1935 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_ofld1()
1936 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_ofld1()
2000 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_destroy() local
2033 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_destroy()
[all …]
Dcnic_if.h369 void (*indicate_kcqes)(void *ulp_ctx, struct kcqe *cqes[],
/Linux-v4.19/drivers/nvme/host/
Dpci.c163 volatile struct nvme_completion *cqes; member
858 return (le16_to_cpu(nvmeq->cqes[nvmeq->cq_head].status) & 1) == in nvme_cqe_pending()
873 volatile struct nvme_completion *cqe = &nvmeq->cqes[idx]; in nvme_handle_cqe()
924 if (nvmeq->cqes[nvmeq->cq_head].command_id == tag) in nvme_process_cq()
1234 (void *)nvmeq->cqes, nvmeq->cq_dma_addr); in nvme_free_queue()
1344 nvmeq->cqes = dma_zalloc_coherent(dev->dev, CQ_SIZE(depth), in nvme_alloc_queue()
1346 if (!nvmeq->cqes) in nvme_alloc_queue()
1367 dma_free_coherent(dev->dev, CQ_SIZE(depth), (void *)nvmeq->cqes, in nvme_alloc_queue()
1396 memset((void *)nvmeq->cqes, 0, CQ_SIZE(nvmeq->q_depth)); in nvme_init_queue()
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/en/
Dxdp.c220 sq->stats->cqes += i; in mlx5e_poll_xdpsq_cq()
/Linux-v4.19/include/linux/
Dnvme.h252 __u8 cqes; member
/Linux-v4.19/drivers/nvme/target/
Dadmin-cmd.c330 id->cqes = (0x4 << 4) | 0x4; in nvmet_execute_identify_ctrl()