/Linux-v4.19/drivers/infiniband/sw/rdmavt/ |
D | cq.c | 79 if (head >= (unsigned)cq->ibcq.cqe) { in rvt_cq_enter() 80 head = cq->ibcq.cqe; in rvt_cq_enter() 88 if (cq->ibcq.event_handler) { in rvt_cq_enter() 91 ev.device = cq->ibcq.device; in rvt_cq_enter() 92 ev.element.cq = &cq->ibcq; in rvt_cq_enter() 94 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rvt_cq_enter() 159 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in send_complete() 278 cq->ibcq.cqe = entries; in rvt_create_cq() 284 ret = &cq->ibcq; in rvt_create_cq() 307 int rvt_destroy_cq(struct ib_cq *ibcq) in rvt_destroy_cq() argument [all …]
|
D | cq.h | 58 int rvt_destroy_cq(struct ib_cq *ibcq); 59 int rvt_req_notify_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags notify_flags); 60 int rvt_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata); 61 int rvt_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *entry);
|
/Linux-v4.19/drivers/infiniband/hw/mlx4/ |
D | cq.c | 44 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_comp() local 45 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx4_ib_cq_comp() 51 struct ib_cq *ibcq; in mlx4_ib_cq_event() local 59 ibcq = &to_mibcq(cq)->ibcq; in mlx4_ib_cq_event() 60 if (ibcq->event_handler) { in mlx4_ib_cq_event() 61 event.device = ibcq->device; in mlx4_ib_cq_event() 63 event.element.cq = ibcq; in mlx4_ib_cq_event() 64 ibcq->event_handler(&event, ibcq->cq_context); in mlx4_ib_cq_event() 80 struct mlx4_cqe *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() 84 !!(n & (cq->ibcq.cqe + 1))) ? NULL : cqe; in get_sw_cqe() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/vmw_pvrdma/ |
D | pvrdma_cq.c | 62 int pvrdma_req_notify_cq(struct ib_cq *ibcq, in pvrdma_req_notify_cq() argument 65 struct pvrdma_dev *dev = to_vdev(ibcq->device); in pvrdma_req_notify_cq() 66 struct pvrdma_cq *cq = to_vcq(ibcq); in pvrdma_req_notify_cq() 82 cq->ibcq.cqe, &head); in pvrdma_req_notify_cq() 135 cq->ibcq.cqe = entries; in pvrdma_create_cq() 199 cq->ibcq.cqe = resp->cqe; in pvrdma_create_cq() 213 pvrdma_destroy_cq(&cq->ibcq); in pvrdma_create_cq() 218 return &cq->ibcq; in pvrdma_create_cq() 299 cq->ibcq.cqe, &head); in _pvrdma_flush_cqe() 304 cq->ibcq.cqe); in _pvrdma_flush_cqe() [all …]
|
D | pvrdma.h | 87 struct ib_cq ibcq; member 274 static inline struct pvrdma_cq *to_vcq(struct ib_cq *ibcq) in to_vcq() argument 276 return container_of(ibcq, struct pvrdma_cq, ibcq); in to_vcq()
|
D | pvrdma_main.c | 354 if (cq && cq->ibcq.event_handler) { in pvrdma_cq_event() 355 struct ib_cq *ibcq = &cq->ibcq; in pvrdma_cq_event() local 358 e.device = ibcq->device; in pvrdma_cq_event() 359 e.element.cq = ibcq; in pvrdma_cq_event() 361 ibcq->event_handler(&e, ibcq->cq_context); in pvrdma_cq_event() 526 if (cq && cq->ibcq.comp_handler) in pvrdma_intrx_handler() 527 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in pvrdma_intrx_handler()
|
/Linux-v4.19/drivers/infiniband/sw/rxe/ |
D | rxe_cq.c | 81 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_send_complete() 113 cq->ibcq.cqe = cqe; in rxe_cq_from_init() 127 cq->ibcq.cqe = cqe; in rxe_cq_resize_queue() 141 if (cq->ibcq.event_handler) { in rxe_cq_post() 142 ev.device = cq->ibcq.device; in rxe_cq_post() 143 ev.element.cq = &cq->ibcq; in rxe_cq_post() 145 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post()
|
D | rxe_verbs.c | 869 return &cq->ibcq; in rxe_create_cq() 877 static int rxe_destroy_cq(struct ib_cq *ibcq) in rxe_destroy_cq() argument 879 struct rxe_cq *cq = to_rcq(ibcq); in rxe_destroy_cq() 887 static int rxe_resize_cq(struct ib_cq *ibcq, int cqe, struct ib_udata *udata) in rxe_resize_cq() argument 890 struct rxe_cq *cq = to_rcq(ibcq); in rxe_resize_cq() 891 struct rxe_dev *rxe = to_rdev(ibcq->device); in rxe_resize_cq() 914 static int rxe_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in rxe_poll_cq() argument 917 struct rxe_cq *cq = to_rcq(ibcq); in rxe_poll_cq() 935 static int rxe_peek_cq(struct ib_cq *ibcq, int wc_cnt) in rxe_peek_cq() argument 937 struct rxe_cq *cq = to_rcq(ibcq); in rxe_peek_cq() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/mthca/ |
D | mthca_cq.c | 181 return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe)); in next_cqe_sw() 235 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 259 event.element.cq = &cq->ibcq; in mthca_cq_event() 260 if (cq->ibcq.event_handler) in mthca_cq_event() 261 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 295 cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe)); in mthca_cq_clean() 297 if (prod_index == cq->cons_index + cq->ibcq.cqe) in mthca_cq_clean() 309 cqe = get_cqe(cq, prod_index & cq->ibcq.cqe); in mthca_cq_clean() 315 memcpy(get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe), in mthca_cq_clean() 321 set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe)); in mthca_cq_clean() [all …]
|
D | mthca_provider.h | 202 struct ib_cq ibcq; member 324 static inline struct mthca_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 326 return container_of(ibcq, struct mthca_cq, ibcq); in to_mcq()
|
/Linux-v4.19/include/rdma/ |
D | rdmavt_cq.h | 82 struct ib_cq ibcq; member 93 static inline struct rvt_cq *ibcq_to_rvtcq(struct ib_cq *ibcq) in ibcq_to_rvtcq() argument 95 return container_of(ibcq, struct rvt_cq, ibcq); in ibcq_to_rvtcq()
|
/Linux-v4.19/drivers/infiniband/hw/mlx5/ |
D | cq.c | 41 struct ib_cq *ibcq = &to_mibcq(cq)->ibcq; in mlx5_ib_cq_comp() local 43 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp() 49 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_ib_cq_event() 50 struct ib_cq *ibcq = &cq->ibcq; in mlx5_ib_cq_event() local 59 if (ibcq->event_handler) { in mlx5_ib_cq_event() 62 event.element.cq = ibcq; in mlx5_ib_cq_event() 63 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event() 79 void *cqe = get_cqe(cq, n & cq->ibcq.cqe); in get_sw_cqe() 85 !((cqe64->op_own & MLX5_CQE_OWNER_MASK) ^ !!(n & (cq->ibcq.cqe + 1)))) { in get_sw_cqe() 512 struct mlx5_ib_dev *dev = to_mdev(cq->ibcq.device); in mlx5_poll_one() [all …]
|
D | mlx5_ib.h | 487 struct ib_cq ibcq; member 926 static inline struct mlx5_ib_cq *to_mcq(struct ib_cq *ibcq) in to_mcq() argument 928 return container_of(ibcq, struct mlx5_ib_cq, ibcq); in to_mcq() 1042 int mlx5_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc); 1043 int mlx5_ib_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags); 1045 int mlx5_ib_resize_cq(struct ib_cq *ibcq, int entries, struct ib_udata *udata); 1108 int mlx5_ib_get_cqe_size(struct mlx5_ib_dev *dev, struct ib_cq *ibcq); 1219 int mlx5_ib_generate_wc(struct ib_cq *ibcq, struct ib_wc *wc);
|
/Linux-v4.19/drivers/infiniband/hw/cxgb3/ |
D | iwch_ev.c | 90 event.device = chp->ibcq.device; in post_qp_event() 92 event.element.cq = &chp->ibcq; in post_qp_event() 100 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 178 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in iwch_ev_dispatch()
|
D | iwch_provider.h | 104 struct ib_cq ibcq; member 114 static inline struct iwch_cq *to_iwch_cq(struct ib_cq *ibcq) in to_iwch_cq() argument 116 return container_of(ibcq, struct iwch_cq, ibcq); in to_iwch_cq() 333 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
D | iwch_qp.c | 748 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in __flush_qp() 762 (*schp->ibcq.comp_handler)(&schp->ibcq, schp->ibcq.cq_context); in __flush_qp() 784 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in flush_qp() 789 (*schp->ibcq.comp_handler)(&schp->ibcq, in flush_qp() 790 schp->ibcq.cq_context); in flush_qp()
|
D | iwch_cq.c | 198 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in iwch_poll_cq() argument 206 chp = to_iwch_cq(ibcq); in iwch_poll_cq()
|
/Linux-v4.19/drivers/infiniband/hw/cxgb4/ |
D | ev.c | 105 event.device = chp->ibcq.device; in post_qp_event() 107 event.element.cq = &chp->ibcq; in post_qp_event() 115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 235 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
|
/Linux-v4.19/drivers/infiniband/hw/hns/ |
D | hns_roce_cq.c | 43 struct ib_cq *ibcq = &hr_cq->ib_cq; in hns_roce_ib_cq_comp() local 45 ibcq->comp_handler(ibcq, ibcq->cq_context); in hns_roce_ib_cq_comp() 53 struct ib_cq *ibcq; in hns_roce_ib_cq_event() local 55 ibcq = &hr_cq->ib_cq; in hns_roce_ib_cq_event() 56 hr_dev = to_hr_dev(ibcq->device); in hns_roce_ib_cq_event() 67 if (ibcq->event_handler) { in hns_roce_ib_cq_event() 68 event.device = ibcq->device; in hns_roce_ib_cq_event() 70 event.element.cq = ibcq; in hns_roce_ib_cq_event() 71 ibcq->event_handler(&event, ibcq->cq_context); in hns_roce_ib_cq_event()
|
/Linux-v4.19/drivers/infiniband/hw/qedr/ |
D | qedr_roce_cm.c | 82 cq->ibcq.comp_handler ? "Yes" : "No"); in qedr_ll2_complete_tx_packet() 92 if (cq->ibcq.comp_handler) in qedr_ll2_complete_tx_packet() 93 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_tx_packet() 121 if (cq->ibcq.comp_handler) in qedr_ll2_complete_rx_packet() 122 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_rx_packet() 669 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in qedr_gsi_poll_cq() argument 671 struct qedr_dev *dev = get_qedr_dev(ibcq->device); in qedr_gsi_poll_cq() 672 struct qedr_cq *cq = get_qedr_cq(ibcq); in qedr_gsi_poll_cq()
|
D | main.c | 496 if (!cq->destroyed && cq->ibcq.comp_handler) in qedr_irq_handler() 497 (*cq->ibcq.comp_handler) in qedr_irq_handler() 498 (&cq->ibcq, cq->ibcq.cq_context); in qedr_irq_handler() 675 struct ib_cq *ibcq; in qedr_affiliated_event() local 734 ibcq = &cq->ibcq; in qedr_affiliated_event() 735 if (ibcq->event_handler) { in qedr_affiliated_event() 736 event.device = ibcq->device; in qedr_affiliated_event() 737 event.element.cq = ibcq; in qedr_affiliated_event() 738 ibcq->event_handler(&event, ibcq->cq_context); in qedr_affiliated_event()
|
D | qedr.h | 274 struct ib_cq ibcq; member 556 static inline struct qedr_cq *get_qedr_cq(struct ib_cq *ibcq) in get_qedr_cq() argument 558 return container_of(ibcq, struct qedr_cq, ibcq); in get_qedr_cq()
|
D | qedr_roce_cm.h | 48 int qedr_gsi_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
|
/Linux-v4.19/drivers/infiniband/hw/i40iw/ |
D | i40iw_hw.c | 175 if (i40iwcq->ibcq.comp_handler) in i40iw_iwarp_ce_handler() 176 i40iwcq->ibcq.comp_handler(&i40iwcq->ibcq, in i40iw_iwarp_ce_handler() 177 i40iwcq->ibcq.cq_context); in i40iw_iwarp_ce_handler() 381 if (iwcq->ibcq.event_handler) { in i40iw_process_aeq() 384 ibevent.device = iwcq->ibcq.device; in i40iw_process_aeq() 386 ibevent.element.cq = &iwcq->ibcq; in i40iw_process_aeq() 387 iwcq->ibcq.event_handler(&ibevent, iwcq->ibcq.cq_context); in i40iw_process_aeq()
|
/Linux-v4.19/drivers/infiniband/hw/ocrdma/ |
D | ocrdma.h | 318 struct ib_cq ibcq; member 473 static inline struct ocrdma_cq *get_ocrdma_cq(struct ib_cq *ibcq) in get_ocrdma_cq() argument 475 return container_of(ibcq, struct ocrdma_cq, ibcq); in get_ocrdma_cq()
|