Home
last modified time | relevance | path

Searched refs:send_cq (Results 1 – 25 of 55) sorted by relevance

123

/Linux-v5.4/drivers/infiniband/ulp/ipoib/
Dipoib_verbs.c187 priv->send_cq = ib_create_cq(priv->ca, ipoib_ib_tx_completion, NULL, in ipoib_transport_dev_init()
189 if (IS_ERR(priv->send_cq)) { in ipoib_transport_dev_init()
197 init_attr.send_cq = priv->send_cq; in ipoib_transport_dev_init()
215 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP)) in ipoib_transport_dev_init()
241 ib_destroy_cq(priv->send_cq); in ipoib_transport_dev_init()
263 ib_destroy_cq(priv->send_cq); in ipoib_transport_dev_cleanup()
Dipoib_ib.c437 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); in poll_tx()
505 n = ib_poll_cq(priv->send_cq, MAX_SEND_CQE, priv->send_wc); in ipoib_tx_poll()
517 if (unlikely(ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_tx_poll()
646 if (ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_send()
Dipoib_cm.c255 .send_cq = priv->recv_cq, /* For drain WR */ in ipoib_cm_create_rx_qp()
769 rc = ib_req_notify_cq(priv->send_cq, IB_CQ_NEXT_COMP | in ipoib_cm_send()
1059 .send_cq = priv->send_cq, in ipoib_cm_create_tx_qp()
/Linux-v5.4/drivers/infiniband/hw/mthca/
Dmthca_qp.c735 qp_context->cqn_snd = cpu_to_be32(to_mcq(ibqp->send_cq)->cqn); in __mthca_modify_qp()
837 if (qp->ibqp.send_cq != qp->ibqp.recv_cq) in __mthca_modify_qp()
838 mthca_cq_clean(dev, to_mcq(qp->ibqp.send_cq), qp->qpn, NULL); in __mthca_modify_qp()
1161 struct mthca_cq *send_cq, in mthca_alloc_qp_common() argument
1290 struct mthca_cq *send_cq, in mthca_alloc_qp() argument
1318 err = mthca_alloc_qp_common(dev, pd, send_cq, recv_cq, in mthca_alloc_qp()
1333 static void mthca_lock_cqs(struct mthca_cq *send_cq, struct mthca_cq *recv_cq) in mthca_lock_cqs() argument
1334 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in mthca_lock_cqs()
1336 if (send_cq == recv_cq) { in mthca_lock_cqs()
1337 spin_lock_irq(&send_cq->lock); in mthca_lock_cqs()
[all …]
Dmthca_dev.h546 struct mthca_cq *send_cq,
555 struct mthca_cq *send_cq,
/Linux-v5.4/include/rdma/
Drdmavt_qp.h821 struct rvt_cq *cq = ibcq_to_rvtcq(qp->ibqp.send_cq); in rvt_send_cq()
936 static inline u32 ib_cq_tail(struct ib_cq *send_cq) in ib_cq_tail() argument
938 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_tail()
940 return ibcq_to_rvtcq(send_cq)->ip ? in ib_cq_tail()
942 ibcq_to_rvtcq(send_cq)->kqueue->tail; in ib_cq_tail()
952 static inline u32 ib_cq_head(struct ib_cq *send_cq) in ib_cq_head() argument
954 struct rvt_cq *cq = ibcq_to_rvtcq(send_cq); in ib_cq_head()
956 return ibcq_to_rvtcq(send_cq)->ip ? in ib_cq_head()
958 ibcq_to_rvtcq(send_cq)->kqueue->head; in ib_cq_head()
/Linux-v5.4/drivers/infiniband/hw/mlx4/
Dqp.c52 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq,
54 static void mlx4_ib_unlock_cqs(struct mlx4_ib_cq *send_cq,
752 if (init_attr->send_cq || init_attr->cap.max_send_wr) { in _mlx4_ib_create_qp_rss()
959 mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), in create_rq()
968 mcq = to_mcq(init_attr->send_cq); in create_rq()
972 mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq), in create_rq()
1246 mlx4_ib_lock_cqs(to_mcq(init_attr->send_cq), in create_qp_common()
1255 mcq = to_mcq(init_attr->send_cq); in create_qp_common()
1259 mlx4_ib_unlock_cqs(to_mcq(init_attr->send_cq), in create_qp_common()
1317 static void mlx4_ib_lock_cqs(struct mlx4_ib_cq *send_cq, struct mlx4_ib_cq *recv_cq) in mlx4_ib_lock_cqs() argument
[all …]
/Linux-v5.4/drivers/infiniband/hw/hns/
Dhns_roce_qp.c1232 void hns_roce_lock_cqs(struct hns_roce_cq *send_cq, struct hns_roce_cq *recv_cq) in hns_roce_lock_cqs() argument
1233 __acquires(&send_cq->lock) __acquires(&recv_cq->lock) in hns_roce_lock_cqs()
1235 if (send_cq == recv_cq) { in hns_roce_lock_cqs()
1236 spin_lock_irq(&send_cq->lock); in hns_roce_lock_cqs()
1238 } else if (send_cq->cqn < recv_cq->cqn) { in hns_roce_lock_cqs()
1239 spin_lock_irq(&send_cq->lock); in hns_roce_lock_cqs()
1243 spin_lock_nested(&send_cq->lock, SINGLE_DEPTH_NESTING); in hns_roce_lock_cqs()
1247 void hns_roce_unlock_cqs(struct hns_roce_cq *send_cq, in hns_roce_unlock_cqs() argument
1248 struct hns_roce_cq *recv_cq) __releases(&send_cq->lock) in hns_roce_unlock_cqs()
1251 if (send_cq == recv_cq) { in hns_roce_unlock_cqs()
[all …]
Dhns_roce_hw_v1.c93 if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { in hns_roce_v1_post_send()
815 hr_qp->ibqp.send_cq = cq; in hns_roce_v1_rsv_lp_qp()
2654 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_sqp()
2697 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_sqp()
2698 hns_roce_v1_cq_clean(to_hr_cq(ibqp->send_cq), in hns_roce_v1_m_sqp()
2800 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_qp()
2866 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_qp()
3311 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v1_m_qp()
3312 hns_roce_v1_cq_clean(to_hr_cq(ibqp->send_cq), in hns_roce_v1_m_qp()
3611 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v1_destroy_qp() local
[all …]
Dhns_roce_hw_v2.c276 if (hns_roce_wq_overflow(&qp->sq, nreq, qp->ibqp.send_cq)) { in hns_roce_v2_post_send()
3531 V2_QPC_BYTE_252_TX_CQN_S, to_hr_cq(ibqp->send_cq)->cqn); in modify_qp_reset_to_init()
3609 V2_QPC_BYTE_252_TX_CQN_S, to_hr_cq(ibqp->send_cq)->cqn); in modify_qp_init_to_init()
4460 if (ibqp->send_cq != ibqp->recv_cq) in hns_roce_v2_modify_qp()
4461 hns_roce_v2_cq_clean(to_hr_cq(ibqp->send_cq), in hns_roce_v2_modify_qp()
4651 struct hns_roce_cq *send_cq, *recv_cq; in hns_roce_v2_destroy_qp_common() local
4665 send_cq = to_hr_cq(hr_qp->ibqp.send_cq); in hns_roce_v2_destroy_qp_common()
4668 hns_roce_lock_cqs(send_cq, recv_cq); in hns_roce_v2_destroy_qp_common()
4673 if (send_cq != recv_cq) in hns_roce_v2_destroy_qp_common()
4674 __hns_roce_v2_cq_clean(send_cq, hr_qp->qpn, NULL); in hns_roce_v2_destroy_qp_common()
[all …]
/Linux-v5.4/drivers/infiniband/hw/mlx5/
Dqp.c101 struct mlx5_ib_cq **send_cq, struct mlx5_ib_cq **recv_cq);
711 static void mlx5_ib_lock_cqs(struct mlx5_ib_cq *send_cq,
713 static void mlx5_ib_unlock_cqs(struct mlx5_ib_cq *send_cq,
1612 if (init_attr->create_flags || init_attr->send_cq) in create_rss_raw_qp_tir()
1882 scqe_sz = mlx5_ib_get_cqe_size(init_attr->send_cq); in configure_requester_scat_cqe()
1955 struct mlx5_ib_cq *send_cq; in create_qp_common() local
2243 if (init_attr->send_cq) in create_qp_common()
2244 MLX5_SET(qpc, qpc, cqn_snd, to_mcq(init_attr->send_cq)->mcq.cqn); in create_qp_common()
2300 get_cqs(init_attr->qp_type, init_attr->send_cq, init_attr->recv_cq, in create_qp_common()
2301 &send_cq, &recv_cq); in create_qp_common()
[all …]
Dgsi.c75 struct ib_cq *gsi_cq = gsi->ibqp.send_cq; in generate_completions()
181 hw_init_attr.send_cq = gsi->cq; in mlx5_ib_gsi_create_qp()
257 .send_cq = gsi->cq, in create_gsi_ud_qp()
/Linux-v5.4/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_qp.c55 static inline void get_cqs(struct pvrdma_qp *qp, struct pvrdma_cq **send_cq, in get_cqs() argument
58 *send_cq = to_vcq(qp->ibqp.send_cq); in get_cqs()
350 cmd->send_cq_handle = to_vcq(init_attr->send_cq)->cq_handle; in pvrdma_create_qp()
989 init_attr->send_cq = qp->ibqp.send_cq; in pvrdma_query_qp()
/Linux-v5.4/drivers/net/ethernet/ibm/ehea/
Dehea_main.c195 arr[i++].fwh = pr->send_cq->fw_handle; in ehea_update_firmware_handles()
800 struct ehea_cq *send_cq = pr->send_cq; in ehea_proc_cqes() local
809 cqe = ehea_poll_cq(send_cq); in ehea_proc_cqes()
811 ehea_inc_cq(send_cq); in ehea_proc_cqes()
851 cqe = ehea_poll_cq(send_cq); in ehea_proc_cqes()
854 ehea_update_feca(send_cq, cqe_counter); in ehea_proc_cqes()
889 ehea_reset_cq_ep(pr->send_cq); in ehea_poll()
891 ehea_reset_cq_n1(pr->send_cq); in ehea_poll()
894 cqe_skb = ehea_poll_cq(pr->send_cq); in ehea_poll()
1473 pr->send_cq = ehea_create_cq(adapter, pr_cfg->max_entries_scq, in ehea_init_port_res()
[all …]
Dehea.h350 struct ehea_cq *send_cq; member
/Linux-v5.4/drivers/infiniband/sw/rxe/
Drxe_qp.c92 if (!init->recv_cq || !init->send_cq) { in rxe_qp_chk_init()
344 struct rxe_cq *scq = to_rcq(init->send_cq); in rxe_qp_from_init()
390 init->send_cq = qp->ibqp.send_cq; in rxe_qp_to_init()
/Linux-v5.4/net/sunrpc/xprtrdma/
Dverbs.c428 ib_free_cq(ep->rep_attr.send_cq); in rpcrdma_ia_remove()
429 ep->rep_attr.send_cq = NULL; in rpcrdma_ia_remove()
541 ep->rep_attr.send_cq = sendcq; in rpcrdma_ep_create()
600 if (ep->rep_attr.send_cq) in rpcrdma_ep_destroy()
601 ib_free_cq(ep->rep_attr.send_cq); in rpcrdma_ep_destroy()
/Linux-v5.4/fs/cifs/
Dsmbdirect.c1464 ib_free_cq(info->send_cq); in smbd_destroy()
1655 info->send_cq = NULL; in _smbd_get_connection()
1657 info->send_cq = in _smbd_get_connection()
1660 if (IS_ERR(info->send_cq)) { in _smbd_get_connection()
1661 info->send_cq = NULL; in _smbd_get_connection()
1683 qp_attr.send_cq = info->send_cq; in _smbd_get_connection()
1802 if (info->send_cq) in _smbd_get_connection()
1803 ib_free_cq(info->send_cq); in _smbd_get_connection()
Dsmbdirect.h59 struct ib_cq *send_cq, *recv_cq; member
/Linux-v5.4/drivers/infiniband/ulp/srp/
Dib_srp.h155 struct ib_cq *send_cq; member
Dib_srp.c544 ib_process_cq_direct(ch->send_cq, -1); in srp_destroy_qp()
556 struct ib_cq *recv_cq, *send_cq; in srp_create_ch_ib() local
575 send_cq = ib_alloc_cq(dev->dev, ch, m * target->queue_size, in srp_create_ch_ib()
577 if (IS_ERR(send_cq)) { in srp_create_ch_ib()
578 ret = PTR_ERR(send_cq); in srp_create_ch_ib()
589 init_attr->send_cq = send_cq; in srp_create_ch_ib()
633 if (ch->send_cq) in srp_create_ch_ib()
634 ib_free_cq(ch->send_cq); in srp_create_ch_ib()
638 ch->send_cq = send_cq; in srp_create_ch_ib()
660 ib_free_cq(send_cq); in srp_create_ch_ib()
[all …]
/Linux-v5.4/drivers/infiniband/core/
Dverbs.c1133 qp->send_cq = qp->recv_cq = NULL; in create_xrc_qp_user()
1218 qp->send_cq = qp_init_attr->send_cq; in ib_create_qp_user()
1222 if (qp_init_attr->send_cq) in ib_create_qp_user()
1223 atomic_inc(&qp_init_attr->send_cq->usecnt); in ib_create_qp_user()
1886 scq = qp->send_cq; in ib_destroy_qp_user()
2642 struct ib_cq *cq = qp->send_cq; in __ib_drain_sq()
/Linux-v5.4/Documentation/infiniband/
Dtag_matching.rst32 processed by the sender. A completion send is received in the send_cq
/Linux-v5.4/drivers/infiniband/hw/qedr/
Dqedr_roce_cm.c62 dev->gsi_sqcq = get_qedr_cq(attrs->send_cq); in qedr_store_gsi_qp_cq()
139 cq = get_qedr_cq(attrs->send_cq); in qedr_destroy_gsi_cq()
/Linux-v5.4/drivers/infiniband/hw/hfi1/
Dqp.c705 ib_cq_head(qp->ibqp.send_cq), in qp_iter_print()
706 ib_cq_tail(qp->ibqp.send_cq), in qp_iter_print()

123