Home
last modified time | relevance | path

Searched refs:ibcq (Results 1 – 25 of 52) sorted by relevance

123

/Linux-v4.19/drivers/infiniband/sw/rdmavt/
Dcq.c79 if (head >= (unsigned)cq->ibcq.cqe) { in rvt_cq_enter()
80 head = cq->ibcq.cqe; in rvt_cq_enter()
88 if (cq->ibcq.event_handler) { in rvt_cq_enter()
91 ev.device = cq->ibcq.device; in rvt_cq_enter()
92 ev.element.cq = &cq->ibcq; in rvt_cq_enter()
94 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rvt_cq_enter()
159 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in send_complete()
278 cq->ibcq.cqe = entries; in rvt_create_cq()
284 ret = &cq->ibcq; in rvt_create_cq()
307 int rvt_destroy_cq(struct ib_cq *ibcq) in rvt_destroy_cq() argument
[all …]
Dcq.h58 int rvt_destroy_cq(struct ib_cq *ibcq);
59 int rvt_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags);
60 int rvt_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata);
61 int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry);
/Linux-v4.19/drivers/infiniband/hw/mlx4/
Dcq.c44 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_comp() local
45 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx4_ib_cq_comp()
51 struct ib_cq *ibcq; in mlx4_ib_cq_event() local
59 ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_event()
60 if (ibcq->event_handler) { in mlx4_ib_cq_event()
61 event.device = ibcq->device; in mlx4_ib_cq_event()
63 event.element.cq = ibcq; in mlx4_ib_cq_event()
64 ibcq->event_handler(&event, ibcq->cq_context); in mlx4_ib_cq_event()
80 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe()
84 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; in get_sw_cqe()
[all …]
/Linux-v4.19/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_cq.c62 int pvrdma_req_notify_cq(struct ib_cq *ibcq, in pvrdma_req_notify_cq() argument
65 struct pvrdma_dev *dev = to_vdev(ibcq->device); in pvrdma_req_notify_cq()
66 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_req_notify_cq()
82 cq->ibcq.cqe, &head); in pvrdma_req_notify_cq()
135 cq->ibcq.cqe = entries; in pvrdma_create_cq()
199 cq->ibcq.cqe = resp->cqe; in pvrdma_create_cq()
213 pvrdma_destroy_cq(&cq->ibcq); in pvrdma_create_cq()
218 return &cq->ibcq; in pvrdma_create_cq()
299 cq->ibcq.cqe, &head); in _pvrdma_flush_cqe()
304 cq->ibcq.cqe); in _pvrdma_flush_cqe()
[all …]
Dpvrdma.h87 struct ib_cq ibcq; member
274 static inline struct pvrdma_cq *to_vcq(struct ib_cq *ibcq) in to_vcq() argument
276 return container_of(ibcq, struct pvrdma_cq, ibcq); in to_vcq()
Dpvrdma_main.c354 if (cq && cq->ibcq.event_handler) { in pvrdma_cq_event()
355 struct ib_cq *ibcq = &cq->ibcq; in pvrdma_cq_event() local
358 e.device = ibcq->device; in pvrdma_cq_event()
359 e.element.cq = ibcq; in pvrdma_cq_event()
361 ibcq->event_handler(&e, ibcq->cq_context); in pvrdma_cq_event()
526 if (cq && cq->ibcq.comp_handler) in pvrdma_intrx_handler()
527 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in pvrdma_intrx_handler()
/Linux-v4.19/drivers/infiniband/sw/rxe/
Drxe_cq.c81 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_send_complete()
113 cq->ibcq.cqe = cqe; in rxe_cq_from_init()
127 cq->ibcq.cqe = cqe; in rxe_cq_resize_queue()
141 if (cq->ibcq.event_handler) { in rxe_cq_post()
142 ev.device = cq->ibcq.device; in rxe_cq_post()
143 ev.element.cq = &cq->ibcq; in rxe_cq_post()
145 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post()
Drxe_verbs.c869 return &cq->ibcq; in rxe_create_cq()
877 static int rxe_destroy_cq(struct ib_cq *ibcq) in rxe_destroy_cq() argument
879 struct rxe_cq *cq = to_rcq(ibcq); in rxe_destroy_cq()
887 static int rxe_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) in rxe_resize_cq() argument
890 struct rxe_cq *cq = to_rcq(ibcq); in rxe_resize_cq()
891 struct rxe_dev *rxe = to_rdev(ibcq->device); in rxe_resize_cq()
914 static int rxe_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in rxe_poll_cq() argument
917 struct rxe_cq *cq = to_rcq(ibcq); in rxe_poll_cq()
935 static int rxe_peek_cq(struct ib_cq *ibcq, int wc_cnt) in rxe_peek_cq() argument
937 struct rxe_cq *cq = to_rcq(ibcq); in rxe_peek_cq()
[all …]
/Linux-v4.19/drivers/infiniband/hw/mthca/
Dmthca_cq.c181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw()
235 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion()
259 event.element.cq = &cq->ibcq; in mthca_cq_event()
260 if (cq->ibcq.event_handler) in mthca_cq_event()
261 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event()
295 cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe)); in mthca_cq_clean()
297 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean()
309 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in mthca_cq_clean()
315 memcpy(get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe), in mthca_cq_clean()
321 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean()
[all …]
Dmthca_provider.h202 struct ib_cq ibcq; member
324 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument
326 return container_of(ibcq, struct mthca_cq, ibcq); in to_mcq()
/Linux-v4.19/include/rdma/
Drdmavt_cq.h82 struct ib_cq ibcq; member
93 static inline struct rvt_cq *ibcq_to_rvtcq(struct ib_cq *ibcq) in ibcq_to_rvtcq() argument
95 return container_of(ibcq, struct rvt_cq, ibcq); in ibcq_to_rvtcq()
/Linux-v4.19/drivers/infiniband/hw/mlx5/
Dcq.c41 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx5_ib_cq_comp() local
43 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp()
49 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_cq_event()
50 struct ib_cq *ibcq = &cq->ibcq; in mlx5_ib_cq_event() local
59 if (ibcq->event_handler) { in mlx5_ib_cq_event()
62 event.element.cq = ibcq; in mlx5_ib_cq_event()
63 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event()
79 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe()
85 !((cqe64->op_own & MLX5_CQE_OWNER_MASK) ^ !!(n & (cq->ibcq.cqe + 1)))) { in get_sw_cqe()
512 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_poll_one()
[all …]
Dmlx5_ib.h487 struct ib_cq ibcq; member
926 static inline struct mlx5_ib_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument
928 return container_of(ibcq, struct mlx5_ib_cq, ibcq); in to_mcq()
1042 int mlx5_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
1043 int mlx5_ib_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags);
1045 int mlx5_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata);
1108 int mlx5_ib_get_cqe_size(struct mlx5_ib_dev *dev, struct ib_cq *ibcq);
1219 int mlx5_ib_generate_wc(struct ib_cq *ibcq, struct ib_wc *wc);
/Linux-v4.19/drivers/infiniband/hw/cxgb3/
Diwch_ev.c90 event.device = chp->ibcq.device; in post_qp_event()
92 event.element.cq = &chp->ibcq; in post_qp_event()
100 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event()
178 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in iwch_ev_dispatch()
Diwch_provider.h104 struct ib_cq ibcq; member
114 static inline struct iwch_cq *to_iwch_cq(struct ib_cq *ibcq) in to_iwch_cq() argument
116 return container_of(ibcq, struct iwch_cq, ibcq); in to_iwch_cq()
333 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
Diwch_qp.c748 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in __flush_qp()
762 (*schp->ibcq.comp_handler)(&schp->ibcq, schp->ibcq.cq_context); in __flush_qp()
784 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in flush_qp()
789 (*schp->ibcq.comp_handler)(&schp->ibcq, in flush_qp()
790 schp->ibcq.cq_context); in flush_qp()
Diwch_cq.c198 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in iwch_poll_cq() argument
206 chp = to_iwch_cq(ibcq); in iwch_poll_cq()
/Linux-v4.19/drivers/infiniband/hw/cxgb4/
Dev.c105 event.device = chp->ibcq.device; in post_qp_event()
107 event.element.cq = &chp->ibcq; in post_qp_event()
115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event()
235 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
/Linux-v4.19/drivers/infiniband/hw/hns/
Dhns_roce_cq.c43 struct ib_cq *ibcq = &hr_cq->ib_cq; in hns_roce_ib_cq_comp() local
45 ibcq->comp_handler(ibcq, ibcq->cq_context); in hns_roce_ib_cq_comp()
53 struct ib_cq *ibcq; in hns_roce_ib_cq_event() local
55 ibcq = &hr_cq->ib_cq; in hns_roce_ib_cq_event()
56 hr_dev = to_hr_dev(ibcq->device); in hns_roce_ib_cq_event()
67 if (ibcq->event_handler) { in hns_roce_ib_cq_event()
68 event.device = ibcq->device; in hns_roce_ib_cq_event()
70 event.element.cq = ibcq; in hns_roce_ib_cq_event()
71 ibcq->event_handler(&event, ibcq->cq_context); in hns_roce_ib_cq_event()
/Linux-v4.19/drivers/infiniband/hw/qedr/
Dqedr_roce_cm.c82 cq->ibcq.comp_handler ? "Yes" : "No"); in qedr_ll2_complete_tx_packet()
92 if (cq->ibcq.comp_handler) in qedr_ll2_complete_tx_packet()
93 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_tx_packet()
121 if (cq->ibcq.comp_handler) in qedr_ll2_complete_rx_packet()
122 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_rx_packet()
669 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in qedr_gsi_poll_cq() argument
671 struct qedr_dev *dev = get_qedr_dev(ibcq->device); in qedr_gsi_poll_cq()
672 struct qedr_cq *cq = get_qedr_cq(ibcq); in qedr_gsi_poll_cq()
Dmain.c496 if (!cq->destroyed && cq->ibcq.comp_handler) in qedr_irq_handler()
497 (*cq->ibcq.comp_handler) in qedr_irq_handler()
498 (&cq->ibcq, cq->ibcq.cq_context); in qedr_irq_handler()
675 struct ib_cq *ibcq; in qedr_affiliated_event() local
734 ibcq = &cq->ibcq; in qedr_affiliated_event()
735 if (ibcq->event_handler) { in qedr_affiliated_event()
736 event.device = ibcq->device; in qedr_affiliated_event()
737 event.element.cq = ibcq; in qedr_affiliated_event()
738 ibcq->event_handler(&event, ibcq->cq_context); in qedr_affiliated_event()
Dqedr.h274 struct ib_cq ibcq; member
556 static inline struct qedr_cq *get_qedr_cq(struct ib_cq *ibcq) in get_qedr_cq() argument
558 return container_of(ibcq, struct qedr_cq, ibcq); in get_qedr_cq()
Dqedr_roce_cm.h48 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
/Linux-v4.19/drivers/infiniband/hw/i40iw/
Di40iw_hw.c175 if (i40iwcq->ibcq.comp_handler) in i40iw_iwarp_ce_handler()
176 i40iwcq->ibcq.comp_handler(&i40iwcq->ibcq, in i40iw_iwarp_ce_handler()
177 i40iwcq->ibcq.cq_context); in i40iw_iwarp_ce_handler()
381 if (iwcq->ibcq.event_handler) { in i40iw_process_aeq()
384 ibevent.device = iwcq->ibcq.device; in i40iw_process_aeq()
386 ibevent.element.cq = &iwcq->ibcq; in i40iw_process_aeq()
387 iwcq->ibcq.event_handler(&ibevent, iwcq->ibcq.cq_context); in i40iw_process_aeq()
/Linux-v4.19/drivers/infiniband/hw/ocrdma/
Docrdma.h318 struct ib_cq ibcq; member
473 static inline struct ocrdma_cq *get_ocrdma_cq(struct ib_cq *ibcq) in get_ocrdma_cq() argument
475 return container_of(ibcq, struct ocrdma_cq, ibcq); in get_ocrdma_cq()

123