| /Linux-v5.4/drivers/infiniband/ulp/ipoib/ |
| D | ipoib_verbs.c | 178 priv->recv_cq = ib_create_cq(priv->ca, ipoib_ib_rx_completion, NULL, in ipoib_transport_dev_init() 180 if (IS_ERR(priv->recv_cq)) { in ipoib_transport_dev_init() 194 if (ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP)) in ipoib_transport_dev_init() 198 init_attr.recv_cq = priv->recv_cq; in ipoib_transport_dev_init() 244 ib_destroy_cq(priv->recv_cq); in ipoib_transport_dev_init() 264 ib_destroy_cq(priv->recv_cq); in ipoib_transport_dev_cleanup()
|
| D | ipoib_ethtool.c | 102 ret = rdma_set_cq_moderation(priv->recv_cq, in ipoib_set_coalesce()
|
| D | ipoib_ib.c | 464 n = ib_poll_cq(priv->recv_cq, t, priv->ibwc); in ipoib_rx_poll() 486 if (unlikely(ib_req_notify_cq(priv->recv_cq, in ipoib_rx_poll() 839 ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP); in ipoib_ib_dev_stop_default() 980 n = ib_poll_cq(priv->recv_cq, IPOIB_NUM_WC, priv->ibwc); in ipoib_drain_cq()
|
| D | ipoib_cm.c | 255 .send_cq = priv->recv_cq, /* For drain WR */ in ipoib_cm_create_rx_qp() 256 .recv_cq = priv->recv_cq, in ipoib_cm_create_rx_qp() 1060 .recv_cq = priv->recv_cq, in ipoib_cm_create_tx_qp()
|
| /Linux-v5.4/drivers/infiniband/hw/mthca/ |
| D | mthca_qp.c | 774 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn); in __mthca_modify_qp() 835 mthca_cq_clean(dev, to_mcq(qp->ibqp.recv_cq), qp->qpn, in __mthca_modify_qp() 837 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) in __mthca_modify_qp() 1162 struct mthca_cq *recv_cq, in mthca_alloc_qp_common() argument 1291 struct mthca_cq *recv_cq, in mthca_alloc_qp() argument 1318 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, in mthca_alloc_qp() 1333 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) in mthca_lock_cqs() argument 1334 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mthca_lock_cqs() 1336 if (send_cq == recv_cq) { in mthca_lock_cqs() 1338 __acquire(&recv_cq->lock); in mthca_lock_cqs() [all …]
|
| D | mthca_dev.h | 547 struct mthca_cq *recv_cq, 556 struct mthca_cq *recv_cq,
|
| /Linux-v5.4/drivers/infiniband/hw/hns/ |
| D | hns_roce_qp.c | 1232 void hns_roce_lock_cqs(struct hns_roce_cq *send_cq, struct hns_roce_cq *recv_cq) in hns_roce_lock_cqs() argument 1233 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in hns_roce_lock_cqs() 1235 if (send_cq == recv_cq) { in hns_roce_lock_cqs() 1237 __acquire(&recv_cq->lock); in hns_roce_lock_cqs() 1238 } else if (send_cq->cqn < recv_cq->cqn) { in hns_roce_lock_cqs() 1240 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING); in hns_roce_lock_cqs() 1242 spin_lock_irq(&recv_cq->lock); in hns_roce_lock_cqs() 1248 struct hns_roce_cq *recv_cq) __releases(&send_cq->lock) in hns_roce_unlock_cqs() argument 1249 __releases(&recv_cq->lock) in hns_roce_unlock_cqs() 1251 if (send_cq == recv_cq) { in hns_roce_unlock_cqs() [all …]
|
| D | hns_roce_hw_v1.c | 370 hr_qp->ibqp.recv_cq)) { in hns_roce_v1_post_recv() 814 hr_qp->ibqp.recv_cq = cq; in hns_roce_v1_rsv_lp_qp() 2650 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_sqp() 2695 hns_roce_v1_cq_clean(to_hr_cq(ibqp->recv_cq), hr_qp->qpn, in hns_roce_v1_m_sqp() 2697 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_sqp() 2804 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_qp() 2870 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_qp() 3309 hns_roce_v1_cq_clean(to_hr_cq(ibqp->recv_cq), hr_qp->qpn, in hns_roce_v1_m_qp() 3311 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_qp() 3611 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v1_destroy_qp() local [all …]
|
| D | hns_roce_hw_v2.c | 642 hr_qp->ibqp.recv_cq)) { in hns_roce_v2_post_recv() 3341 V2_QPC_BYTE_80_RX_CQN_S, to_hr_cq(ibqp->recv_cq)->cqn); in modify_qp_reset_to_init() 3604 V2_QPC_BYTE_80_RX_CQN_S, to_hr_cq(ibqp->recv_cq)->cqn); in modify_qp_init_to_init() 4458 hns_roce_v2_cq_clean(to_hr_cq(ibqp->recv_cq), hr_qp->qpn, in hns_roce_v2_modify_qp() 4460 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v2_modify_qp() 4651 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v2_destroy_qp_common() local 4666 recv_cq = to_hr_cq(hr_qp->ibqp.recv_cq); in hns_roce_v2_destroy_qp_common() 4668 hns_roce_lock_cqs(send_cq, recv_cq); in hns_roce_v2_destroy_qp_common() 4671 __hns_roce_v2_cq_clean(recv_cq, hr_qp->qpn, hr_qp->ibqp.srq ? in hns_roce_v2_destroy_qp_common() 4673 if (send_cq != recv_cq) in hns_roce_v2_destroy_qp_common() [all …]
|
| /Linux-v5.4/drivers/infiniband/hw/mlx4/ |
| D | qp.c | 53 struct mlx4_ib_cq *recv_cq); 55 struct mlx4_ib_cq *recv_cq); 960 to_mcq(init_attr->recv_cq)); in create_rq() 970 mcq = to_mcq(init_attr->recv_cq); in create_rq() 973 to_mcq(init_attr->recv_cq)); in create_rq() 1247 to_mcq(init_attr->recv_cq)); in create_qp_common() 1257 mcq = to_mcq(init_attr->recv_cq); in create_qp_common() 1260 to_mcq(init_attr->recv_cq)); in create_qp_common() 1317 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) in mlx4_ib_lock_cqs() argument 1318 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mlx4_ib_lock_cqs() [all …]
|
| /Linux-v5.4/drivers/infiniband/hw/mlx5/ |
| D | qp.c | 101 struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq); 712 struct mlx5_ib_cq *recv_cq); 714 struct mlx5_ib_cq *recv_cq); 1850 rcqe_sz = mlx5_ib_get_cqe_size(init_attr->recv_cq); in configure_responder_scat_cqe() 1956 struct mlx5_ib_cq *recv_cq; in create_qp_common() local 2246 if (init_attr->recv_cq) in create_qp_common() 2247 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(init_attr->recv_cq)->mcq.cqn); in create_qp_common() 2300 get_cqs(init_attr->qp_type, init_attr->send_cq, init_attr->recv_cq, in create_qp_common() 2301 &send_cq, &recv_cq); in create_qp_common() 2303 mlx5_ib_lock_cqs(send_cq, recv_cq); in create_qp_common() [all …]
|
| D | gsi.c | 258 .recv_cq = gsi->rx_qp->recv_cq, in create_gsi_ud_qp()
|
| /Linux-v5.4/drivers/infiniband/hw/vmw_pvrdma/ |
| D | pvrdma_qp.c | 56 struct pvrdma_cq **recv_cq) in get_cqs() argument 59 *recv_cq = to_vcq(qp->ibqp.recv_cq); in get_cqs() 351 cmd->recv_cq_handle = to_vcq(init_attr->recv_cq)->cq_handle; in pvrdma_create_qp() 990 init_attr->recv_cq = qp->ibqp.recv_cq; in pvrdma_query_qp()
|
| /Linux-v5.4/drivers/infiniband/sw/rxe/ |
| D | rxe_qp.c | 92 if (!init->recv_cq || !init->send_cq) { in rxe_qp_chk_init() 343 struct rxe_cq *rcq = to_rcq(init->recv_cq); in rxe_qp_from_init() 391 init->recv_cq = qp->ibqp.recv_cq; in rxe_qp_to_init()
|
| /Linux-v5.4/net/sunrpc/xprtrdma/ |
| D | verbs.c | 426 ib_free_cq(ep->rep_attr.recv_cq); in rpcrdma_ia_remove() 427 ep->rep_attr.recv_cq = NULL; in rpcrdma_ia_remove() 542 ep->rep_attr.recv_cq = recvcq; in rpcrdma_ep_create() 598 if (ep->rep_attr.recv_cq) in rpcrdma_ep_destroy() 599 ib_free_cq(ep->rep_attr.recv_cq); in rpcrdma_ep_destroy()
|
| /Linux-v5.4/fs/cifs/ |
| D | smbdirect.c | 1465 ib_free_cq(info->recv_cq); in smbd_destroy() 1656 info->recv_cq = NULL; in _smbd_get_connection() 1665 info->recv_cq = in _smbd_get_connection() 1668 if (IS_ERR(info->recv_cq)) { in _smbd_get_connection() 1669 info->recv_cq = NULL; in _smbd_get_connection() 1684 qp_attr.recv_cq = info->recv_cq; in _smbd_get_connection() 1804 if (info->recv_cq) in _smbd_get_connection() 1805 ib_free_cq(info->recv_cq); in _smbd_get_connection()
|
| D | smbdirect.h | 59 struct ib_cq *send_cq, *recv_cq; member
|
| /Linux-v5.4/drivers/infiniband/core/ |
| D | verbs.c | 1133 qp->send_cq = qp->recv_cq = NULL; in create_xrc_qp_user() 1157 (qp_init_attr->recv_cq || in ib_create_qp_user() 1207 qp->recv_cq = NULL; in ib_create_qp_user() 1210 qp->recv_cq = qp_init_attr->recv_cq; in ib_create_qp_user() 1211 if (qp_init_attr->recv_cq) in ib_create_qp_user() 1212 atomic_inc(&qp_init_attr->recv_cq->usecnt); in ib_create_qp_user() 1887 rcq = qp->recv_cq; in ib_destroy_qp_user() 2681 struct ib_cq *cq = qp->recv_cq; in __ib_drain_rq()
|
| /Linux-v5.4/drivers/infiniband/ulp/srp/ |
| D | ib_srp.h | 156 struct ib_cq *recv_cq; member
|
| D | ib_srp.c | 556 struct ib_cq *recv_cq, *send_cq; in srp_create_ch_ib() local 568 recv_cq = ib_alloc_cq(dev->dev, ch, target->queue_size + 1, in srp_create_ch_ib() 570 if (IS_ERR(recv_cq)) { in srp_create_ch_ib() 571 ret = PTR_ERR(recv_cq); in srp_create_ch_ib() 590 init_attr->recv_cq = recv_cq; in srp_create_ch_ib() 631 if (ch->recv_cq) in srp_create_ch_ib() 632 ib_free_cq(ch->recv_cq); in srp_create_ch_ib() 637 ch->recv_cq = recv_cq; in srp_create_ch_ib() 663 ib_free_cq(recv_cq); in srp_create_ch_ib() 709 ib_free_cq(ch->recv_cq); in srp_free_ch_ib() [all …]
|
| /Linux-v5.4/drivers/net/ethernet/ibm/ehea/ |
| D | ehea_main.c | 197 arr[i++].fwh = pr->recv_cq->fw_handle; in ehea_update_firmware_handles() 888 ehea_reset_cq_ep(pr->recv_cq); in ehea_poll() 890 ehea_reset_cq_n1(pr->recv_cq); in ehea_poll() 1465 pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq, in ehea_init_port_res() 1468 if (!pr->recv_cq) { in ehea_init_port_res() 1484 pr->recv_cq->attr.act_nr_of_cqes); in ehea_init_port_res() 1509 init_attr->recv_cq_handle = pr->recv_cq->fw_handle; in ehea_init_port_res() 1559 ehea_destroy_cq(pr->recv_cq); in ehea_init_port_res() 1576 ehea_destroy_cq(pr->recv_cq); in ehea_clean_portres()
|
| D | ehea.h | 351 struct ehea_cq *recv_cq; member
|
| /Linux-v5.4/drivers/infiniband/hw/qedr/ |
| D | qedr_roce_cm.c | 63 dev->gsi_rqcq = get_qedr_cq(attrs->recv_cq); in qedr_store_gsi_qp_cq() 144 cq = get_qedr_cq(attrs->recv_cq); in qedr_destroy_gsi_cq()
|
| /Linux-v5.4/drivers/infiniband/sw/rdmavt/ |
| D | qp.c | 1342 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, 1); in rvt_error_qp() 1372 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, 1); in rvt_error_qp() 1781 init_attr->recv_cq = qp->ibqp.recv_cq; in rvt_query_qp() 1845 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, 1); in rvt_post_recv() 2331 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, 1); in init_sge()
|
| /Linux-v5.4/drivers/infiniband/sw/siw/ |
| D | siw_verbs.c | 357 rcq = to_siw_cq(attrs->recv_cq); in siw_create_qp() 550 qp_init_attr->recv_cq = base_qp->recv_cq; in siw_query_qp()
|