Home
last modified time | relevance | path

Searched refs:cqes (Results 1 – 17 of 17) sorted by relevance

/Linux-v5.10/tools/io_uring/
Dsetup.c41 cq->ring_sz = p->cq_off.cqes + p->cq_entries * sizeof(struct io_uring_cqe); in io_uring_mmap()
54 cq->cqes = ptr + p->cq_off.cqes; in io_uring_mmap()
Dio_uring-bench.c51 struct io_uring_cqe *cqes; member
257 cqe = &ring->cqes[head & cq_ring_mask]; in reap_events()
449 ptr = mmap(0, p.cq_off.cqes + p.cq_entries * sizeof(struct io_uring_cqe), in setup_ring()
457 cring->cqes = ptr + p.cq_off.cqes; in setup_ring()
Dqueue.c31 *cqe_ptr = &cq->cqes[head & mask]; in __io_uring_get_cqe()
Dliburing.h41 struct io_uring_cqe *cqes; member
/Linux-v5.10/drivers/infiniband/hw/cxgb4/
Drestrack.c334 struct t4_cqe *cqes) in fill_hwcqes() argument
339 if (fill_cqe(msg, cqes, idx, "hwcq_idx")) in fill_hwcqes()
342 if (fill_cqe(msg, cqes + 1, idx, "hwcq_idx")) in fill_hwcqes()
351 struct t4_cqe *cqes) in fill_swcqes() argument
359 if (fill_cqe(msg, cqes, idx, "swcq_idx")) in fill_swcqes()
364 if (fill_cqe(msg, cqes + 1, idx, "swcq_idx")) in fill_swcqes()
/Linux-v5.10/drivers/net/ethernet/broadcom/
Dcnic.c1425 struct kcqe *cqes[], u32 num_cqes) in cnic_reply_bnx2x_kcqes() argument
1434 cqes, num_cqes); in cnic_reply_bnx2x_kcqes()
1552 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_init2() local
1582 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_init2()
1583 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_init2()
1881 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_ofld1() local
1934 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_ofld1()
1935 cnic_reply_bnx2x_kcqes(dev, CNIC_ULP_ISCSI, cqes, 1); in cnic_bnx2x_iscsi_ofld1()
1999 struct kcqe *cqes[1]; in cnic_bnx2x_iscsi_destroy() local
2032 cqes[0] = (struct kcqe *) &kcqe; in cnic_bnx2x_iscsi_destroy()
[all …]
Dcnic_if.h369 void (*indicate_kcqes)(void *ulp_ctx, struct kcqe *cqes[],
/Linux-v5.10/drivers/net/ethernet/mellanox/mlx5/core/
Den_stats.c289 s->rx_xdp_tx_cqe += xdpsq_stats->cqes; in MLX5E_DECLARE_STATS_GRP_OP_UPDATE_STATS()
331 s->tx_xdp_cqes += xdpsq_red_stats->cqes; in MLX5E_DECLARE_STATS_GRP_OP_UPDATE_STATS()
357 s->tx_xsk_cqes += xsksq_stats->cqes; in MLX5E_DECLARE_STATS_GRP_OP_UPDATE_STATS()
393 s->tx_cqes += sq_stats->cqes; in MLX5E_DECLARE_STATS_GRP_OP_UPDATE_STATS()
1594 { MLX5E_DECLARE_TX_STAT(struct mlx5e_sq_stats, cqes) },
1606 { MLX5E_DECLARE_RQ_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) },
1616 { MLX5E_DECLARE_XDPSQ_STAT(struct mlx5e_xdpsq_stats, cqes) },
1647 { MLX5E_DECLARE_XSKSQ_STAT(struct mlx5e_xdpsq_stats, cqes) },
Den_stats.h376 u64 cqes ____cacheline_aligned_in_smp;
389 u64 cqes ____cacheline_aligned_in_smp;
Den_tx.c816 stats->cqes += i; in mlx5e_poll_tx_cq()
/Linux-v5.10/include/uapi/linux/
Dio_uring.h210 __u32 cqes; member
/Linux-v5.10/drivers/nvme/target/
Dpassthru.c77 id->cqes = min_t(__u8, ((0x4 << 4) | 0x4), id->cqes); in nvmet_passthru_override_id_ctrl()
Dadmin-cmd.c405 id->cqes = (0x4 << 4) | 0x4; in nvmet_execute_identify_ctrl()
/Linux-v5.10/drivers/nvme/host/
Dpci.c194 struct nvme_completion *cqes; member
946 struct nvme_completion *hcqe = &nvmeq->cqes[nvmeq->cq_head]; in nvme_cqe_pending()
969 struct nvme_completion *cqe = &nvmeq->cqes[idx]; in nvme_handle_cqe()
1343 (void *)nvmeq->cqes, nvmeq->cq_dma_addr); in nvme_free_queue()
1483 nvmeq->cqes = dma_alloc_coherent(dev->dev, CQ_SIZE(nvmeq), in nvme_alloc_queue()
1485 if (!nvmeq->cqes) in nvme_alloc_queue()
1503 dma_free_coherent(dev->dev, CQ_SIZE(nvmeq), (void *)nvmeq->cqes, in nvme_alloc_queue()
1532 memset((void *)nvmeq->cqes, 0, CQ_SIZE(nvmeq)); in nvme_init_queue()
/Linux-v5.10/drivers/net/ethernet/mellanox/mlx5/core/en/
Dxdp.c453 sq->stats->cqes += i; in mlx5e_poll_xdpsq_cq()
/Linux-v5.10/include/linux/
Dnvme.h280 __u8 cqes; member
/Linux-v5.10/fs/
Dio_uring.c187 struct io_uring_cqe cqes[] ____cacheline_aligned_in_smp; member
1564 return &rings->cqes[tail & ctx->cq_mask]; in io_get_cqring()
7943 off = struct_size(rings, cqes, cq_entries); in rings_size()
9365 p->cq_off.cqes = offsetof(struct io_rings, cqes); in io_uring_create()