Home
last modified time | relevance | path

Searched defs:cqe (Results 26 – 50 of 166) sorted by relevance

1234567

/Linux-v6.6/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_cq.c294 struct pvrdma_cqe *cqe; in _pvrdma_flush_cqe() local
329 struct pvrdma_cqe *cqe; in pvrdma_poll_one() local
/Linux-v6.6/drivers/net/ethernet/mellanox/mlx5/core/
Dwq.h202 struct mlx5_cqe64 *cqe = mlx5_frag_buf_get_wqe(&wq->fbc, ix); in mlx5_cqwq_get_wqe() local
233 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(wq, ci); in mlx5_cqwq_get_cqe() local
251 struct mlx5_cqe64 *cqe; in mlx5_cqwq_get_cqe_enahnced_comp() local
Den_tc.h372 static inline bool mlx5e_cqe_regb_chain(struct mlx5_cqe64 *cqe) in mlx5e_cqe_regb_chain()
397 static inline bool mlx5e_cqe_regb_chain(struct mlx5_cqe64 *cqe) in mlx5e_cqe_regb_chain()
400 mlx5e_tc_update_skb_nic(struct mlx5_cqe64 *cqe, struct sk_buff *skb) in mlx5e_tc_update_skb_nic()
/Linux-v6.6/drivers/infiniband/hw/erdma/
Derdma_cmdq.c257 __be32 *cqe = get_queue_entry(cmdq->cq.qbuf, cmdq->cq.ci, in get_next_valid_cmdq_cqe() local
293 __be32 *cqe; in erdma_poll_single_cmd_completion() local
/Linux-v6.6/drivers/infiniband/sw/siw/
Dsiw_cq.c50 struct siw_cqe *cqe; in siw_reap_cqe() local
Dsiw_qp.c1071 struct siw_cqe *cqe; in siw_sqe_complete() local
1128 struct siw_cqe *cqe; in siw_rqe_complete() local
/Linux-v6.6/drivers/infiniband/ulp/rtrs/
Drtrs.c97 int rtrs_post_recv_empty(struct rtrs_con *con, struct ib_cqe *cqe) in rtrs_post_recv_empty()
188 struct ib_cqe *cqe, in rtrs_post_rdma_write_imm_empty()
386 void rtrs_init_hb(struct rtrs_path *path, struct ib_cqe *cqe, in rtrs_init_hb()
/Linux-v6.6/drivers/net/ethernet/mellanox/mlx4/
Den_rx.c625 static int check_csum(struct mlx4_cqe *cqe, struct sk_buff *skb, void *va, in check_csum()
666 struct mlx4_cqe *cqe; member
688 struct mlx4_cqe *cqe = _ctx->cqe; in mlx4_en_xdp_rx_hash() local
723 struct mlx4_cqe *cqe; in mlx4_en_process_rx_cq() local
/Linux-v6.6/drivers/net/ethernet/mellanox/mlx5/core/en/rep/
Dtc.h68 mlx5e_rep_tc_receive(struct mlx5_cqe64 *cqe, struct mlx5e_rq *rq, in mlx5e_rep_tc_receive()
/Linux-v6.6/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_qp.c842 struct hinic_rq_cqe *cqe; in hinic_rq_read_wqe() local
902 struct hinic_rq_cqe *cqe = rq->cqe[cons_idx]; in hinic_rq_put_wqe() local
925 struct hinic_rq_cqe *cqe = rq->cqe[cons_idx]; in hinic_rq_get_sge() local
945 struct hinic_rq_cqe *cqe = rq->cqe[prod_idx]; in hinic_rq_prepare_wqe() local
Dhinic_rx.c45 #define LRO_PKT_HDR_LEN(cqe) \ argument
363 struct hinic_rq_cqe *cqe; in rxq_recv() local
/Linux-v6.6/drivers/infiniband/hw/mlx5/
Dumr.h80 struct ib_cqe cqe; member
/Linux-v6.6/drivers/infiniband/hw/bnxt_re/
Dib_verbs.c2927 int cqe = attr->cqe; in bnxt_re_create_cq() local
3043 int bnxt_re_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) in bnxt_re_resize_cq()
3212 static void bnxt_re_process_req_wc(struct ib_wc *wc, struct bnxt_qplib_cqe *cqe) in bnxt_re_process_req_wc()
3346 struct bnxt_qplib_cqe *cqe) in bnxt_re_process_raw_qp_pkt_rx()
3457 struct bnxt_qplib_cqe *cqe) in bnxt_re_process_res_rawqp1_wc()
3504 struct bnxt_qplib_cqe *cqe) in bnxt_re_process_res_rc_wc()
3520 struct bnxt_qplib_cqe *cqe) in bnxt_re_process_res_shadow_qp_wc()
3568 struct bnxt_qplib_cqe *cqe) in bnxt_re_process_res_ud_wc()
3627 struct bnxt_qplib_cqe *cqe; in bnxt_re_poll_cq() local
/Linux-v6.6/drivers/net/ethernet/mellanox/mlx5/core/en_accel/
Dktls_txrx.h87 struct mlx5_cqe64 *cqe, in mlx5e_ktls_handle_rx_skb()
/Linux-v6.6/drivers/net/ethernet/mellanox/mlx5/core/lib/
Daso.c65 struct mlx5_cqe64 *cqe = mlx5_cqwq_get_wqe(&cq->wq, i); in mlx5_aso_alloc_cq() local
394 struct mlx5_cqe64 *cqe; in mlx5_aso_poll_cq() local
/Linux-v6.6/drivers/scsi/elx/efct/
Defct_hw.c361 efct_hw_wq_process_io(void *arg, u8 *cqe, int status) in efct_hw_wq_process_io()
1847 efct_hw_wq_process_abort(void *arg, u8 *cqe, int status) in efct_hw_wq_process_abort()
2091 void (*callback)(void *arg, u8 *cqe, int status), in efct_hw_reqtag_alloc()
2315 u8 cqe[sizeof(struct sli4_mcqe)]; in efct_hw_cq_process() local
2413 u8 *cqe, int status, u16 rid) in efct_hw_wq_process()
2440 u8 *cqe, u16 rid) in efct_hw_xabt_process()
2972 void (*callback)(void *arg, u8 *cqe, int status), in efct_hw_send_frame()
/Linux-v6.6/drivers/net/ethernet/ibm/ehea/
Dehea_main.c532 static inline int ehea_check_cqe(struct ehea_cqe *cqe, int *rq_num) in ehea_check_cqe()
544 struct sk_buff *skb, struct ehea_cqe *cqe, in ehea_fill_skb()
565 struct ehea_cqe *cqe) in get_skb_by_index()
618 struct ehea_cqe *cqe, int *processed_rq2, in ehea_treat_poll_error()
659 struct ehea_cqe *cqe; in ehea_proc_rwqes() local
805 struct ehea_cqe *cqe; in ehea_proc_cqes() local
882 struct ehea_cqe *cqe; in ehea_poll() local
/Linux-v6.6/include/linux/mlx5/
Ddevice.h882 static inline u8 mlx5_get_cqe_format(struct mlx5_cqe64 *cqe) in mlx5_get_cqe_format()
887 static inline u8 get_cqe_opcode(struct mlx5_cqe64 *cqe) in get_cqe_opcode()
892 static inline u8 get_cqe_enhanced_num_mini_cqes(struct mlx5_cqe64 *cqe) in get_cqe_enhanced_num_mini_cqes()
898 static inline u8 get_cqe_lro_tcppsh(struct mlx5_cqe64 *cqe) in get_cqe_lro_tcppsh()
903 static inline u8 get_cqe_l4_hdr_type(struct mlx5_cqe64 *cqe) in get_cqe_l4_hdr_type()
908 static inline bool cqe_is_tunneled(struct mlx5_cqe64 *cqe) in cqe_is_tunneled()
913 static inline u8 get_cqe_tls_offload(struct mlx5_cqe64 *cqe) in get_cqe_tls_offload()
918 static inline bool cqe_has_vlan(struct mlx5_cqe64 *cqe) in cqe_has_vlan()
923 static inline u64 get_cqe_ts(struct mlx5_cqe64 *cqe) in get_cqe_ts()
933 static inline u16 get_cqe_flow_tag(struct mlx5_cqe64 *cqe) in get_cqe_flow_tag()
[all …]
/Linux-v6.6/drivers/vfio/pci/mlx5/
Dcmd.c877 void *cqe; in init_cq_frag_buf() local
1499 mlx5vf_rq_cqe(struct mlx5_vhca_qp *qp, struct mlx5_cqe64 *cqe, in mlx5vf_rq_cqe()
1526 void *cqe = get_cqe(cq, n & (cq->ncqe - 1)); in get_sw_cqe() local
1543 struct mlx5_cqe64 *cqe; in mlx5vf_cq_poll_one() local
/Linux-v6.6/net/sunrpc/xprtrdma/
Dsvc_rdma_rw.c266 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_write_done() local
333 struct ib_cqe *cqe = wc->wr_cqe; in svc_rdma_wc_read_done() local
369 struct ib_cqe *cqe; in svc_rdma_post_chunk_ctxt() local
/Linux-v6.6/drivers/infiniband/hw/ocrdma/
Docrdma_hw.c123 struct ocrdma_mcqe *cqe = (struct ocrdma_mcqe *) in ocrdma_get_mcqe() local
676 struct ocrdma_ae_mcqe *cqe) in ocrdma_dispatch_ibevent()
802 struct ocrdma_ae_mcqe *cqe) in ocrdma_process_grp5_aync()
828 struct ocrdma_ae_mcqe *cqe) in ocrdma_process_link_state()
846 struct ocrdma_ae_mcqe *cqe = ae_cqe; in ocrdma_process_acqe() local
865 static void ocrdma_process_mcqe(struct ocrdma_dev *dev, struct ocrdma_mcqe *cqe) in ocrdma_process_mcqe()
883 struct ocrdma_mcqe *cqe; in ocrdma_mq_cq_handler() local
/Linux-v6.6/drivers/infiniband/core/
Drw.c498 u32 port_num, struct ib_cqe *cqe, struct ib_send_wr *chain_wr) in rdma_rw_ctx_wrs()
556 struct ib_cqe *cqe, struct ib_send_wr *chain_wr) in rdma_rw_ctx_post()
/Linux-v6.6/drivers/scsi/qedf/
Dqedf_els.c142 void qedf_process_els_compl(struct qedf_ctx *qedf, struct fcoe_cqe *cqe, in qedf_process_els_compl()
744 struct fcoe_cqe *cqe, struct qedf_ioreq *io_req) in qedf_process_seq_cleanup_compl()
/Linux-v6.6/drivers/nvme/target/
Dloop.c21 struct nvme_completion cqe; member
95 struct nvme_completion *cqe = req->cqe; in nvme_loop_queue_response() local
/Linux-v6.6/drivers/ufs/core/
Dufs-mcq.c263 struct cq_entry *cqe) in ufshcd_mcq_get_tag()
280 struct cq_entry *cqe = ufshcd_mcq_cur_cqe(hwq); in ufshcd_mcq_process_cqe() local

1234567