Home
last modified time | relevance | path

Searched refs:cqp (Results 1 – 19 of 19) sorted by relevance

/Linux-v5.15/drivers/infiniband/hw/irdma/
Dctrl.c156 irdma_sc_add_arp_cache_entry(struct irdma_sc_cqp *cqp, in irdma_sc_add_arp_cache_entry() argument
163 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_add_arp_cache_entry()
173 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_sc_add_arp_cache_entry()
181 irdma_sc_cqp_post_sq(cqp); in irdma_sc_add_arp_cache_entry()
194 irdma_sc_del_arp_cache_entry(struct irdma_sc_cqp *cqp, u64 scratch, in irdma_sc_del_arp_cache_entry() argument
200 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_del_arp_cache_entry()
206 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_sc_del_arp_cache_entry()
215 irdma_sc_cqp_post_sq(cqp); in irdma_sc_del_arp_cache_entry()
228 irdma_sc_manage_apbvt_entry(struct irdma_sc_cqp *cqp, in irdma_sc_manage_apbvt_entry() argument
235 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_manage_apbvt_entry()
[all …]
Duda.h39 enum irdma_status_code irdma_sc_access_ah(struct irdma_sc_cqp *cqp, struct irdma_ah_info *info,
41 enum irdma_status_code irdma_access_mcast_grp(struct irdma_sc_cqp *cqp,
50 static inline enum irdma_status_code irdma_sc_create_ah(struct irdma_sc_cqp *cqp, in irdma_sc_create_ah() argument
54 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_CREATE_ADDR_HANDLE, in irdma_sc_create_ah()
58 static inline enum irdma_status_code irdma_sc_destroy_ah(struct irdma_sc_cqp *cqp, in irdma_sc_destroy_ah() argument
62 return irdma_sc_access_ah(cqp, info, IRDMA_CQP_OP_DESTROY_ADDR_HANDLE, in irdma_sc_destroy_ah()
66 static inline enum irdma_status_code irdma_sc_create_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_create_mcast_grp() argument
70 return irdma_access_mcast_grp(cqp, info, IRDMA_CQP_OP_CREATE_MCAST_GRP, in irdma_sc_create_mcast_grp()
74 static inline enum irdma_status_code irdma_sc_modify_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_sc_modify_mcast_grp() argument
78 return irdma_access_mcast_grp(cqp, info, IRDMA_CQP_OP_MODIFY_MCAST_GRP, in irdma_sc_modify_mcast_grp()
[all …]
Dutils.c420 struct irdma_cqp_request *irdma_alloc_and_get_cqp_request(struct irdma_cqp *cqp, in irdma_alloc_and_get_cqp_request() argument
426 spin_lock_irqsave(&cqp->req_lock, flags); in irdma_alloc_and_get_cqp_request()
427 if (!list_empty(&cqp->cqp_avail_reqs)) { in irdma_alloc_and_get_cqp_request()
428 cqp_request = list_first_entry(&cqp->cqp_avail_reqs, in irdma_alloc_and_get_cqp_request()
432 spin_unlock_irqrestore(&cqp->req_lock, flags); in irdma_alloc_and_get_cqp_request()
442 ibdev_dbg(to_ibdev(cqp->sc_cqp.dev), "ERR: CQP Request Fail: No Memory"); in irdma_alloc_and_get_cqp_request()
467 void irdma_free_cqp_request(struct irdma_cqp *cqp, in irdma_free_cqp_request() argument
479 spin_lock_irqsave(&cqp->req_lock, flags); in irdma_free_cqp_request()
480 list_add_tail(&cqp_request->list, &cqp->cqp_avail_reqs); in irdma_free_cqp_request()
481 spin_unlock_irqrestore(&cqp->req_lock, flags); in irdma_free_cqp_request()
[all …]
Duda.c19 enum irdma_status_code irdma_sc_access_ah(struct irdma_sc_cqp *cqp, in irdma_sc_access_ah() argument
26 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_sc_access_ah()
69 FIELD_PREP(IRDMA_UDA_CQPSQ_MAV_WQEVALID, cqp->polarity) | in irdma_sc_access_ah()
78 irdma_sc_cqp_post_sq(cqp); in irdma_sc_access_ah()
118 enum irdma_status_code irdma_access_mcast_grp(struct irdma_sc_cqp *cqp, in irdma_access_mcast_grp() argument
126 ibdev_dbg(to_ibdev(cqp->dev), "WQE: mg_id out of range\n"); in irdma_access_mcast_grp()
130 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, scratch); in irdma_access_mcast_grp()
132 ibdev_dbg(to_ibdev(cqp->dev), "WQE: ring full\n"); in irdma_access_mcast_grp()
163 FIELD_PREP(IRDMA_UDA_CQPSQ_MG_WQEVALID, cqp->polarity) | in irdma_access_mcast_grp()
174 irdma_sc_cqp_post_sq(cqp); in irdma_access_mcast_grp()
Dhw.c564 struct irdma_cqp *cqp = &rf->cqp; in irdma_destroy_cqp() local
569 status = irdma_sc_cqp_destroy(dev->cqp); in irdma_destroy_cqp()
574 dma_free_coherent(dev->hw->device, cqp->sq.size, cqp->sq.va, in irdma_destroy_cqp()
575 cqp->sq.pa); in irdma_destroy_cqp()
576 cqp->sq.va = NULL; in irdma_destroy_cqp()
577 kfree(cqp->scratch_array); in irdma_destroy_cqp()
578 cqp->scratch_array = NULL; in irdma_destroy_cqp()
579 kfree(cqp->cqp_requests); in irdma_destroy_cqp()
580 cqp->cqp_requests = NULL; in irdma_destroy_cqp()
840 return irdma_sc_static_hmc_pages_allocated(dev->cqp, 0, dev->hmc_fn_id, in irdma_create_hmc_objs()
[all …]
Dtype.h644 struct irdma_sc_cqp *cqp; member
708 struct irdma_sc_cqp *cqp; member
1232 enum irdma_status_code irdma_sc_cqp_create(struct irdma_sc_cqp *cqp, u16 *maj_err,
1234 enum irdma_status_code irdma_sc_cqp_destroy(struct irdma_sc_cqp *cqp);
1235 enum irdma_status_code irdma_sc_cqp_init(struct irdma_sc_cqp *cqp,
1237 void irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp);
1238 enum irdma_status_code irdma_sc_poll_for_cqp_op_done(struct irdma_sc_cqp *cqp, u8 opcode,
1270 enum irdma_status_code irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp,
1338 struct irdma_sc_cqp *cqp; member
1344 struct irdma_sc_cqp *cqp; member
[all …]
Dprotos.h19 void irdma_sc_cqp_post_sq(struct irdma_sc_cqp *cqp);
20 __le64 *irdma_sc_cqp_get_next_send_wqe(struct irdma_sc_cqp *cqp, u64 scratch);
87 irdma_sc_static_hmc_pages_allocated(struct irdma_sc_cqp *cqp, u64 scratch,
Dpuda.c609 struct irdma_sc_cqp *cqp; in irdma_puda_qp_wqe() local
615 cqp = dev->cqp; in irdma_puda_qp_wqe()
616 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, 0); in irdma_puda_qp_wqe()
628 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_puda_qp_wqe()
635 irdma_sc_cqp_post_sq(cqp); in irdma_puda_qp_wqe()
636 status = irdma_sc_poll_for_cqp_op_done(dev->cqp, IRDMA_CQP_OP_CREATE_QP, in irdma_puda_qp_wqe()
729 struct irdma_sc_cqp *cqp; in irdma_puda_cq_wqe() local
734 cqp = dev->cqp; in irdma_puda_cq_wqe()
735 wqe = irdma_sc_cqp_get_next_send_wqe(cqp, 0); in irdma_puda_cq_wqe()
754 FIELD_PREP(IRDMA_CQPSQ_WQEVALID, cqp->polarity); in irdma_puda_cq_wqe()
[all …]
Dmain.h300 struct irdma_cqp cqp; member
475 struct irdma_cqp_request *irdma_alloc_and_get_cqp_request(struct irdma_cqp *cqp,
477 void irdma_free_cqp_request(struct irdma_cqp *cqp,
479 void irdma_put_cqp_request(struct irdma_cqp *cqp,
Dverbs.c260 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_alloc_push_page()
272 cqp_info->in.u.manage_push_page.cqp = &iwdev->rf->cqp.sc_cqp; in irdma_alloc_push_page()
282 irdma_put_cqp_request(&iwdev->rf->cqp, cqp_request); in irdma_alloc_push_page()
672 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_cqp_create_qp_cmd()
688 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_cqp_create_qp_cmd()
1868 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_resize_cq()
1886 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_resize_cq()
2083 cqp_request = irdma_alloc_and_get_cqp_request(&rf->cqp, true); in irdma_create_cq()
2096 irdma_put_cqp_request(&rf->cqp, cqp_request); in irdma_create_cq()
2436 cqp_request = irdma_alloc_and_get_cqp_request(&iwdev->rf->cqp, true); in irdma_hw_alloc_mw()
[all …]
Dhmc.c137 return dev->cqp->process_cqp_sds(dev, &sdinfo); in irdma_hmc_sd_one()
177 ret_code = dev->cqp->process_cqp_sds(dev, &sdinfo); in irdma_hmc_sd_grp()
189 ret_code = dev->cqp->process_cqp_sds(dev, &sdinfo); in irdma_hmc_sd_grp()
/Linux-v5.15/drivers/net/ethernet/mellanox/mlx5/core/en/
Dparams.h27 struct mlx5e_cq_param cqp; member
34 struct mlx5e_cq_param cqp; member
Dparams.c541 mlx5e_build_rx_cq_param(mdev, params, xsk, &param->cqp); in mlx5e_build_rq_param()
600 mlx5e_build_tx_cq_param(mdev, params, &param->cqp); in mlx5e_build_sq_param()
655 mlx5e_build_ico_cq_param(mdev, log_wq_size, &param->cqp); in mlx5e_build_icosq_param()
672 mlx5e_build_ico_cq_param(mdev, log_wq_size, &param->cqp); in mlx5e_build_async_icosq_param()
685 mlx5e_build_tx_cq_param(mdev, params, &param->cqp); in mlx5e_build_xdpsq_param()
Dptp.c378 cq_param = &cparams->txq_sq_param.cqp; in mlx5e_ptp_open_tx_cqs()
425 cq_param = &cparams->rq_param.cqp; in mlx5e_ptp_open_rx_cq()
453 mlx5e_build_tx_cq_param(mdev, params, &param->cqp); in mlx5e_ptp_build_sq_param()
Dtrap.c69 err = mlx5e_open_cq(priv, trap_moder, &rq_param->cqp, &ccp, &rq->cq); in mlx5e_open_trap_rq()
/Linux-v5.15/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/
Dsetup.c112 err = mlx5e_open_cq(c->priv, params->rx_cq_moderation, &cparam->rq.cqp, &ccp, in mlx5e_open_xsk()
121 err = mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->xdp_sq.cqp, &ccp, in mlx5e_open_xsk()
/Linux-v5.15/drivers/scsi/lpfc/
Dlpfc_sli4.h1077 int lpfc_cq_create_set(struct lpfc_hba *phba, struct lpfc_queue **cqp,
1087 struct lpfc_queue **drqp, struct lpfc_queue **cqp,
Dlpfc_sli.c16346 lpfc_cq_create_set(struct lpfc_hba *phba, struct lpfc_queue **cqp, in lpfc_cq_create_set() argument
16363 if (!cqp || !hdwq || !numcq) in lpfc_cq_create_set()
16371 length += ((numcq * cqp[0]->page_count) * in lpfc_cq_create_set()
16389 cq = cqp[idx]; in lpfc_cq_create_set()
16563 cq = cqp[idx]; in lpfc_cq_create_set()
17365 struct lpfc_queue **drqp, struct lpfc_queue **cqp, in lpfc_mrq_create() argument
17380 if (!hrqp || !drqp || !cqp || !numrq) in lpfc_mrq_create()
17416 cq = cqp[idx]; in lpfc_mrq_create()
/Linux-v5.15/drivers/net/ethernet/mellanox/mlx5/core/
Den_main.c1685 err = mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->txq_sq.cqp, in mlx5e_open_tx_cqs()
1843 err = mlx5e_open_cq(c->priv, icocq_moder, &cparam->async_icosq.cqp, &ccp, in mlx5e_open_queues()
1848 err = mlx5e_open_cq(c->priv, icocq_moder, &cparam->icosq.cqp, &ccp, in mlx5e_open_queues()
1857 err = mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->xdp_sq.cqp, &ccp, in mlx5e_open_queues()
1862 err = mlx5e_open_cq(c->priv, params->rx_cq_moderation, &cparam->rq.cqp, &ccp, in mlx5e_open_queues()
1867 err = c->xdp ? mlx5e_open_cq(c->priv, params->tx_cq_moderation, &cparam->xdp_sq.cqp, in mlx5e_open_queues()