Home
last modified time | relevance | path

Searched refs:recv_cq (Results 1 – 25 of 60) sorted by relevance

123

/Linux-v4.19/drivers/infiniband/ulp/ipoib/
Dipoib_verbs.c178 priv->recv_cq = ib_create_cq(priv->ca, ipoib_ib_rx_completion, NULL, in ipoib_transport_dev_init()
180 if (IS_ERR(priv->recv_cq)) { in ipoib_transport_dev_init()
194 if (ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP)) in ipoib_transport_dev_init()
198 init_attr.recv_cq = priv->recv_cq; in ipoib_transport_dev_init()
244 ib_destroy_cq(priv->recv_cq); in ipoib_transport_dev_init()
266 if (ib_destroy_cq(priv->recv_cq)) in ipoib_transport_dev_cleanup()
Dipoib_ethtool.c102 ret = rdma_set_cq_moderation(priv->recv_cq, in ipoib_set_coalesce()
Dipoib_ib.c463 n = ib_poll_cq(priv->recv_cq, t, priv->ibwc); in ipoib_rx_poll()
485 if (unlikely(ib_req_notify_cq(priv->recv_cq, in ipoib_rx_poll()
839 ib_req_notify_cq(priv->recv_cq, IB_CQ_NEXT_COMP); in ipoib_ib_dev_stop_default()
980 n = ib_poll_cq(priv->recv_cq, IPOIB_NUM_WC, priv->ibwc); in ipoib_drain_cq()
Dipoib_cm.c255 .send_cq = priv->recv_cq, /* For drain WR */ in ipoib_cm_create_rx_qp()
256 .recv_cq = priv->recv_cq, in ipoib_cm_create_rx_qp()
1060 .recv_cq = priv->recv_cq, in ipoib_cm_create_tx_qp()
/Linux-v4.19/drivers/infiniband/hw/mthca/
Dmthca_qp.c770 qp_context->cqn_rcv = cpu_to_be32(to_mcq(ibqp->recv_cq)->cqn); in __mthca_modify_qp()
831 mthca_cq_clean(dev, to_mcq(qp->ibqp.recv_cq), qp->qpn, in __mthca_modify_qp()
833 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) in __mthca_modify_qp()
1156 struct mthca_cq *recv_cq, in mthca_alloc_qp_common() argument
1284 struct mthca_cq *recv_cq, in mthca_alloc_qp() argument
1310 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, in mthca_alloc_qp()
1325 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) in mthca_lock_cqs() argument
1326 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mthca_lock_cqs()
1328 if (send_cq == recv_cq) { in mthca_lock_cqs()
1330 __acquire(&recv_cq->lock); in mthca_lock_cqs()
[all …]
Dmthca_dev.h546 struct mthca_cq *recv_cq,
554 struct mthca_cq *recv_cq,
/Linux-v4.19/drivers/infiniband/hw/hns/
Dhns_roce_qp.c1028 void hns_roce_lock_cqs(struct hns_roce_cq *send_cq, struct hns_roce_cq *recv_cq) in hns_roce_lock_cqs() argument
1029 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in hns_roce_lock_cqs()
1031 if (send_cq == recv_cq) { in hns_roce_lock_cqs()
1033 __acquire(&recv_cq->lock); in hns_roce_lock_cqs()
1034 } else if (send_cq->cqn < recv_cq->cqn) { in hns_roce_lock_cqs()
1036 spin_lock_nested(&recv_cq->lock, SINGLE_DEPTH_NESTING); in hns_roce_lock_cqs()
1038 spin_lock_irq(&recv_cq->lock); in hns_roce_lock_cqs()
1045 struct hns_roce_cq *recv_cq) __releases(&send_cq->lock) in hns_roce_unlock_cqs() argument
1046 __releases(&recv_cq->lock) in hns_roce_unlock_cqs()
1048 if (send_cq == recv_cq) { in hns_roce_unlock_cqs()
[all …]
Dhns_roce_hw_v1.c373 hr_qp->ibqp.recv_cq)) { in hns_roce_v1_post_recv()
806 hr_qp->ibqp.recv_cq = cq; in hns_roce_v1_rsv_lp_qp()
2686 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_sqp()
2731 hns_roce_v1_cq_clean(to_hr_cq(ibqp->recv_cq), hr_qp->qpn, in hns_roce_v1_m_sqp()
2733 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_sqp()
2840 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_qp()
2906 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_qp()
3345 hns_roce_v1_cq_clean(to_hr_cq(ibqp->recv_cq), hr_qp->qpn, in hns_roce_v1_m_qp()
3347 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_qp()
3928 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v1_destroy_qp() local
[all …]
Dhns_roce_hw_v2.c583 hr_qp->ibqp.recv_cq)) { in hns_roce_v2_post_recv()
2638 V2_QPC_BYTE_80_RX_CQN_S, to_hr_cq(ibqp->recv_cq)->cqn); in modify_qp_reset_to_init()
2921 V2_QPC_BYTE_80_RX_CQN_S, to_hr_cq(ibqp->recv_cq)->cqn); in modify_qp_init_to_init()
3635 hns_roce_v2_cq_clean(to_hr_cq(ibqp->recv_cq), hr_qp->qpn, in hns_roce_v2_modify_qp()
3637 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v2_modify_qp()
3833 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v2_destroy_qp_common() local
3849 recv_cq = to_hr_cq(hr_qp->ibqp.recv_cq); in hns_roce_v2_destroy_qp_common()
3851 hns_roce_lock_cqs(send_cq, recv_cq); in hns_roce_v2_destroy_qp_common()
3854 __hns_roce_v2_cq_clean(recv_cq, hr_qp->qpn, hr_qp->ibqp.srq ? in hns_roce_v2_destroy_qp_common()
3856 if (send_cq != recv_cq) in hns_roce_v2_destroy_qp_common()
[all …]
Dhns_roce_device.h997 struct hns_roce_cq *recv_cq);
999 struct hns_roce_cq *recv_cq);
/Linux-v4.19/drivers/infiniband/hw/mlx5/
Dqp.c98 struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq);
629 struct mlx5_ib_cq *recv_cq);
631 struct mlx5_ib_cq *recv_cq);
1612 struct mlx5_ib_cq *recv_cq; in create_qp_common() local
1817 rcqe_sz = mlx5_ib_get_cqe_size(dev, init_attr->recv_cq); in create_qp_common()
1876 if (init_attr->recv_cq) in create_qp_common()
1877 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(init_attr->recv_cq)->mcq.cqn); in create_qp_common()
1929 get_cqs(init_attr->qp_type, init_attr->send_cq, init_attr->recv_cq, in create_qp_common()
1930 &send_cq, &recv_cq); in create_qp_common()
1932 mlx5_ib_lock_cqs(send_cq, recv_cq); in create_qp_common()
[all …]
Dgsi.c258 .recv_cq = gsi->rx_qp->recv_cq, in create_gsi_ud_qp()
/Linux-v4.19/drivers/infiniband/hw/mlx4/
Dqp.c52 struct mlx4_ib_cq *recv_cq);
54 struct mlx4_ib_cq *recv_cq);
1161 to_mcq(init_attr->recv_cq)); in create_qp_common()
1171 mcq = to_mcq(init_attr->recv_cq); in create_qp_common()
1174 to_mcq(init_attr->recv_cq)); in create_qp_common()
1235 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) in mlx4_ib_lock_cqs() argument
1236 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mlx4_ib_lock_cqs()
1238 if (send_cq == recv_cq) { in mlx4_ib_lock_cqs()
1240 __acquire(&recv_cq->lock); in mlx4_ib_lock_cqs()
1241 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx4_ib_lock_cqs()
[all …]
/Linux-v4.19/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_qp.c56 struct pvrdma_cq **recv_cq) in get_cqs() argument
59 *recv_cq = to_vcq(qp->ibqp.recv_cq); in get_cqs()
353 cmd->recv_cq_handle = to_vcq(init_attr->recv_cq)->cq_handle; in pvrdma_create_qp()
993 init_attr->recv_cq = qp->ibqp.recv_cq; in pvrdma_query_qp()
/Linux-v4.19/drivers/infiniband/core/
Dverbs.c1098 qp->send_cq = qp->recv_cq = NULL; in ib_create_xrc_qp()
1121 (qp_init_attr->recv_cq || in ib_create_qp()
1162 qp->recv_cq = NULL; in ib_create_qp()
1165 qp->recv_cq = qp_init_attr->recv_cq; in ib_create_qp()
1166 if (qp_init_attr->recv_cq) in ib_create_qp()
1167 atomic_inc(&qp_init_attr->recv_cq->usecnt); in ib_create_qp()
1832 rcq = qp->recv_cq; in ib_destroy_qp()
2517 struct ib_cq *cq = qp->recv_cq; in __ib_drain_rq()
/Linux-v4.19/drivers/infiniband/sw/rxe/
Drxe_qp.c90 if (!init->recv_cq || !init->send_cq) { in rxe_qp_chk_init()
331 struct rxe_cq *rcq = to_rcq(init->recv_cq); in rxe_qp_from_init()
380 init->recv_cq = qp->ibqp.recv_cq; in rxe_qp_to_init()
/Linux-v4.19/net/sunrpc/xprtrdma/
Dverbs.c451 ib_free_cq(ep->rep_attr.recv_cq); in rpcrdma_ia_remove()
452 ep->rep_attr.recv_cq = NULL; in rpcrdma_ia_remove()
568 ep->rep_attr.recv_cq = recvcq; in rpcrdma_ep_create()
626 if (ep->rep_attr.recv_cq) in rpcrdma_ep_destroy()
627 ib_free_cq(ep->rep_attr.recv_cq); in rpcrdma_ep_destroy()
/Linux-v4.19/drivers/infiniband/ulp/srp/
Dib_srp.h143 struct ib_cq *recv_cq; member
Dib_srp.c546 struct ib_cq *recv_cq, *send_cq; in srp_create_ch_ib() local
558 recv_cq = ib_alloc_cq(dev->dev, ch, target->queue_size + 1, in srp_create_ch_ib()
560 if (IS_ERR(recv_cq)) { in srp_create_ch_ib()
561 ret = PTR_ERR(recv_cq); in srp_create_ch_ib()
580 init_attr->recv_cq = recv_cq; in srp_create_ch_ib()
621 if (ch->recv_cq) in srp_create_ch_ib()
622 ib_free_cq(ch->recv_cq); in srp_create_ch_ib()
627 ch->recv_cq = recv_cq; in srp_create_ch_ib()
653 ib_free_cq(recv_cq); in srp_create_ch_ib()
699 ib_free_cq(ch->recv_cq); in srp_free_ch_ib()
[all …]
/Linux-v4.19/fs/cifs/
Dsmbdirect.c235 ib_free_cq(info->recv_cq); in smbd_destroy_rdma_work()
1679 info->recv_cq = NULL; in _smbd_get_connection()
1687 info->recv_cq = ib_alloc_cq(info->id->device, info, in _smbd_get_connection()
1689 if (IS_ERR(info->recv_cq)) { in _smbd_get_connection()
1690 info->recv_cq = NULL; in _smbd_get_connection()
1705 qp_attr.recv_cq = info->recv_cq; in _smbd_get_connection()
1833 if (info->recv_cq) in _smbd_get_connection()
1834 ib_free_cq(info->recv_cq); in _smbd_get_connection()
Dsmbdirect.h68 struct ib_cq *send_cq, *recv_cq; member
/Linux-v4.19/drivers/infiniband/hw/qib/
Dqib_ud.c218 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, in qib_ud_loopback()
581 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, in qib_ud_rcv()
/Linux-v4.19/drivers/net/ethernet/ibm/ehea/
Dehea_main.c211 arr[i++].fwh = pr->recv_cq->fw_handle; in ehea_update_firmware_handles()
903 ehea_reset_cq_ep(pr->recv_cq); in ehea_poll()
905 ehea_reset_cq_n1(pr->recv_cq); in ehea_poll()
1480 pr->recv_cq = ehea_create_cq(adapter, pr_cfg->max_entries_rcq, in ehea_init_port_res()
1483 if (!pr->recv_cq) { in ehea_init_port_res()
1499 pr->recv_cq->attr.act_nr_of_cqes); in ehea_init_port_res()
1524 init_attr->recv_cq_handle = pr->recv_cq->fw_handle; in ehea_init_port_res()
1574 ehea_destroy_cq(pr->recv_cq); in ehea_init_port_res()
1591 ehea_destroy_cq(pr->recv_cq); in ehea_clean_portres()
Dehea.h365 struct ehea_cq *recv_cq; member
/Linux-v4.19/drivers/infiniband/sw/rdmavt/
Dqp.c1086 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, 1); in rvt_error_qp()
1109 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, 1); in rvt_error_qp()
1519 init_attr->recv_cq = qp->ibqp.recv_cq; in rvt_query_qp()
1583 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, 1); in rvt_post_recv()
2039 rvt_cq_enter(ibcq_to_rvtcq(qp->ibqp.recv_cq), &wc, 1); in init_sge()

123