Searched refs:cqes (Results 1 – 10 of 10) sorted by relevance
/Linux-v4.19/drivers/infiniband/hw/cxgb4/ |
D | restrack.c | 337 struct t4_cqe *cqes) in fill_hwcqes() argument 342 if (fill_cqe(msg, cqes, idx, "hwcq_idx")) in fill_hwcqes() 345 if (fill_cqe(msg, cqes + 1, idx, "hwcq_idx")) in fill_hwcqes() 354 struct t4_cqe *cqes) in fill_swcqes() argument 362 if (fill_cqe(msg, cqes, idx, "swcq_idx")) in fill_swcqes() 367 if (fill_cqe(msg, cqes + 1, idx, "swcq_idx")) in fill_swcqes()
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_stats.h | 229 u64 cqes ____cacheline_aligned_in_smp; 239 u64 cqes ____cacheline_aligned_in_smp;
|
D | en_stats.c | 157 s->rx_xdp_tx_cqe += xdpsq_stats->cqes; in mlx5e_grp_sw_update_stats() 180 s->tx_xdp_cqes += xdpsq_red_stats->cqes; in mlx5e_grp_sw_update_stats() 207 s->tx_cqes += sq_stats->cqes; in mlx5e_grp_sw_update_stats() 1179 { MLX5E_DECLARE_TX_STAT(struct mlx5e_sq_stats, cqes) }, 1188 { MLX5E_DECLARE_RQ_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) }, 1195 { MLX5E_DECLARE_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) },
|
D | en_tx.c | 563 stats->cqes += i; in mlx5e_poll_tx_cq()
|
/Linux-v4.19/drivers/net/ethernet/broadcom/ |
D | cnic.c | 1425 struct kcqe *cqes[], u32 num_cqes) in cnic_reply_bnx2x_kcqes() argument 1434 cqes, num_cqes); in cnic_reply_bnx2x_kcqes() 1552 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_init2() local 1582 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_init2() 1583 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_init2() 1881 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_ofld1() local 1935 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_ofld1() 1936 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_ofld1() 2000 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_destroy() local 2033 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_destroy() [all …]
|
D | cnic_if.h | 369 void (*indicate_kcqes)(void *ulp_ctx, struct kcqe *cqes[],
|
/Linux-v4.19/drivers/nvme/host/ |
D | pci.c | 163 volatile struct nvme_completion *cqes; member 858 return (le16_to_cpu(nvmeq->cqes[nvmeq->cq_head].status) & 1) == in nvme_cqe_pending() 873 volatile struct nvme_completion *cqe = &nvmeq->cqes[idx]; in nvme_handle_cqe() 924 if (nvmeq->cqes[nvmeq->cq_head].command_id == tag) in nvme_process_cq() 1234 (void *)nvmeq->cqes, nvmeq->cq_dma_addr); in nvme_free_queue() 1344 nvmeq->cqes = dma_zalloc_coherent(dev->dev, CQ_SIZE(depth), in nvme_alloc_queue() 1346 if (!nvmeq->cqes) in nvme_alloc_queue() 1367 dma_free_coherent(dev->dev, CQ_SIZE(depth), (void *)nvmeq->cqes, in nvme_alloc_queue() 1396 memset((void *)nvmeq->cqes, 0, CQ_SIZE(nvmeq->q_depth)); in nvme_init_queue()
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | xdp.c | 220 sq->stats->cqes += i; in mlx5e_poll_xdpsq_cq()
|
/Linux-v4.19/include/linux/ |
D | nvme.h | 252 __u8 cqes; member
|
/Linux-v4.19/drivers/nvme/target/ |
D | admin-cmd.c | 330 id->cqes = (0x4 << 4) | 0x4; in nvmet_execute_identify_ctrl()
|