/Linux-v5.15/drivers/infiniband/hw/mlx5/ |
D | gsi.c | 47 static void generate_completions(struct mlx5_ib_qp *mqp) in generate_completions() argument 49 struct mlx5_ib_gsi_qp *gsi = &mqp->gsi; in generate_completions() 50 struct ib_cq *gsi_cq = mqp->ibqp.send_cq; in generate_completions() 73 struct mlx5_ib_qp *mqp = container_of(gsi, struct mlx5_ib_qp, gsi); in handle_single_completion() local 82 wr->wc.qp = &mqp->ibqp; in handle_single_completion() 84 generate_completions(mqp); in handle_single_completion() 88 int mlx5_ib_create_gsi(struct ib_pd *pd, struct mlx5_ib_qp *mqp, in mlx5_ib_create_gsi() argument 106 gsi = &mqp->gsi; in mlx5_ib_create_gsi() 168 int mlx5_ib_destroy_gsi(struct mlx5_ib_qp *mqp) in mlx5_ib_destroy_gsi() argument 170 struct mlx5_ib_dev *dev = to_mdev(mqp->ibqp.device); in mlx5_ib_destroy_gsi() [all …]
|
D | qp.c | 1322 err = mlx5_core_create_sq_tracked(dev, in, inlen, &sq->base.mqp); in create_raw_packet_qp_sq() 1342 mlx5_core_destroy_sq_tracked(dev, &sq->base.mqp); in destroy_raw_packet_qp_sq() 1350 struct mlx5_ib_qp *mqp = rq->base.container_mibqp; in create_raw_packet_qp_rq() local 1391 if (mqp->flags & IB_QP_CREATE_SCATTER_FCS) in create_raw_packet_qp_rq() 1409 err = mlx5_core_create_rq_tracked(dev, in, inlen, &rq->base.mqp); in create_raw_packet_qp_rq() 1419 mlx5_core_destroy_rq_tracked(dev, &rq->base.mqp); in destroy_raw_packet_qp_rq() 1452 MLX5_SET(tirc, tirc, inline_rqn, rq->base.mqp.qpn); in create_raw_packet_qp_tir() 1514 resp->sqn = sq->base.mqp.qpn; in create_raw_packet_qp() 1519 sq->base.mqp.event = mlx5_ib_qp_event; in create_raw_packet_qp() 1540 resp->rqn = rq->base.mqp.qpn; in create_raw_packet_qp() [all …]
|
D | qpc.c | 193 struct mlx5_core_qp *qp = &dct->mqp; in _mlx5_core_destroy_dct() 206 destroy_resource_common(dev, &dct->mqp); in _mlx5_core_destroy_dct() 217 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_create_dct() 275 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_drain_dct() 520 struct mlx5_core_qp *qp = &dct->mqp; in mlx5_core_dct_query()
|
D | fs.c | 1151 struct mlx5_ib_qp *mqp = to_mqp(qp); in mlx5_ib_create_flow() local 1237 if (mqp->is_rss) in mlx5_ib_create_flow() 1238 dst->tir_num = mqp->rss_qp.tirn; in mlx5_ib_create_flow() 1240 dst->tir_num = mqp->raw_packet_qp.rq.tirn; in mlx5_ib_create_flow() 1245 underlay_qpn = (mqp->flags & IB_QP_CREATE_SOURCE_QPN) ? in mlx5_ib_create_flow() 1246 mqp->underlay_qpn : in mlx5_ib_create_flow() 1871 struct mlx5_ib_qp *mqp; in get_dests() local 1881 mqp = to_mqp(*qp); in get_dests() 1882 if (mqp->is_rss) in get_dests() 1883 *dest_id = mqp->rss_qp.tirn; in get_dests() [all …]
|
D | mlx5_ib.h | 403 struct mlx5_core_qp mqp; member 1136 static inline struct mlx5_ib_qp *to_mibqp(struct mlx5_core_qp *mqp) in to_mibqp() argument 1138 return container_of(mqp, struct mlx5_ib_qp_base, mqp)->container_mibqp; in to_mibqp() 1405 int mlx5_ib_create_gsi(struct ib_pd *pd, struct mlx5_ib_qp *mqp, 1407 int mlx5_ib_destroy_gsi(struct mlx5_ib_qp *mqp);
|
D | main.c | 2395 struct mlx5_ib_qp *mqp = to_mqp(ibqp); in mlx5_ib_mcg_attach() local 2402 if (mqp->flags & IB_QP_CREATE_SOURCE_QPN) { in mlx5_ib_mcg_attach() 2526 struct mlx5_ib_qp *mqp; in mlx5_ib_handle_internal_error() local 2538 list_for_each_entry(mqp, &ibdev->qp_list, qps_list) { in mlx5_ib_handle_internal_error() 2539 spin_lock_irqsave(&mqp->sq.lock, flags_qp); in mlx5_ib_handle_internal_error() 2540 if (mqp->sq.tail != mqp->sq.head) { in mlx5_ib_handle_internal_error() 2541 send_mcq = to_mcq(mqp->ibqp.send_cq); in mlx5_ib_handle_internal_error() 2544 mqp->ibqp.send_cq->comp_handler) { in mlx5_ib_handle_internal_error() 2553 spin_unlock_irqrestore(&mqp->sq.lock, flags_qp); in mlx5_ib_handle_internal_error() 2554 spin_lock_irqsave(&mqp->rq.lock, flags_qp); in mlx5_ib_handle_internal_error() [all …]
|
D | ib_rep.c | 189 return mlx5_eswitch_add_send_to_vport_rule(esw, esw, rep, sq->base.mqp.qpn); in create_flow_rule_vport_sq()
|
D | mem.c | 128 (qp->trans_qp.base.mqp.qpn << 8)); in post_send_nop()
|
D | cq.c | 448 struct mlx5_core_qp *mqp; in mlx5_poll_one() local 489 mqp = radix_tree_lookup(&dev->qp_table.tree, qpn); in mlx5_poll_one() 490 *cur_qp = to_mibqp(mqp); in mlx5_poll_one()
|
D | odp.c | 1077 u32 qpn = qp->trans_qp.base.mqp.qpn; in mlx5_ib_mr_initiator_pfault_handler() 1195 struct mlx5_core_qp *mqp = (struct mlx5_core_qp *)res; in res_to_qp() local 1197 return to_mibqp(mqp); in res_to_qp()
|
D | devx.c | 642 rq->base.mqp.qpn) == obj_id || in devx_is_valid_obj_id() 644 sq->base.mqp.qpn) == obj_id || in devx_is_valid_obj_id() 653 qp->dct.mdct.mqp.qpn) == obj_id; in devx_is_valid_obj_id()
|
D | wr.c | 990 ctrl->qpn_ds = cpu_to_be32(size | (qp->trans_qp.base.mqp.qpn << 8)); in finish_wqe()
|
/Linux-v5.15/drivers/infiniband/sw/rdmavt/ |
D | mcast.c | 35 struct rvt_mcast_qp *mqp; in rvt_mcast_qp_alloc() local 37 mqp = kmalloc(sizeof(*mqp), GFP_KERNEL); in rvt_mcast_qp_alloc() 38 if (!mqp) in rvt_mcast_qp_alloc() 41 mqp->qp = qp; in rvt_mcast_qp_alloc() 45 return mqp; in rvt_mcast_qp_alloc() 48 static void rvt_mcast_qp_free(struct rvt_mcast_qp *mqp) in rvt_mcast_qp_free() argument 50 struct rvt_qp *qp = mqp->qp; in rvt_mcast_qp_free() 55 kfree(mqp); in rvt_mcast_qp_free() 152 struct rvt_mcast *mcast, struct rvt_mcast_qp *mqp) in rvt_mcast_add() argument 186 if (p->qp == mqp->qp) { in rvt_mcast_add() [all …]
|
/Linux-v5.15/drivers/scsi/arm/ |
D | msgqueue.c | 121 struct msgqueue_entry **mqp; in msgqueue_addmsg() local 133 mqp = &msgq->qe; in msgqueue_addmsg() 134 while (*mqp) in msgqueue_addmsg() 135 mqp = &(*mqp)->next; in msgqueue_addmsg() 137 *mqp = mq; in msgqueue_addmsg()
|
/Linux-v5.15/drivers/infiniband/hw/mlx4/ |
D | main.c | 1270 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in add_gid_entry() local 1279 if (mlx4_ib_add_mc(mdev, mqp, gid)) { in add_gid_entry() 1280 ge->port = mqp->port; in add_gid_entry() 1284 mutex_lock(&mqp->mutex); in add_gid_entry() 1285 list_add_tail(&ge->list, &mqp->gid_list); in add_gid_entry() 1286 mutex_unlock(&mqp->mutex); in add_gid_entry() 1307 int mlx4_ib_add_mc(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp, in mlx4_ib_add_mc() argument 1313 if (!mqp->port) in mlx4_ib_add_mc() 1317 ndev = mdev->iboe.netdevs[mqp->port - 1]; in mlx4_ib_add_mc() 1845 struct mlx4_ib_qp *mqp = to_mqp(ibqp); in mlx4_ib_mcg_attach() local [all …]
|
D | qp.c | 110 return qp->mqp.qpn >= dev->dev->phys_caps.base_tunnel_sqpn && in is_tunnel_qp() 111 qp->mqp.qpn < dev->dev->phys_caps.base_tunnel_sqpn + in is_tunnel_qp() 122 qp->mqp.qpn >= dev->dev->phys_caps.base_sqpn && in is_sqp() 123 qp->mqp.qpn <= dev->dev->phys_caps.base_sqpn + 3); in is_sqp() 129 if (qp->mqp.qpn == dev->dev->caps.spec_qps[i].qp0_proxy || in is_sqp() 130 qp->mqp.qpn == dev->dev->caps.spec_qps[i].qp1_proxy) { in is_sqp() 150 qp->mqp.qpn >= dev->dev->phys_caps.base_sqpn && in is_qp0() 151 qp->mqp.qpn <= dev->dev->phys_caps.base_sqpn + 1); in is_qp0() 157 if (qp->mqp.qpn == dev->dev->caps.spec_qps[i].qp0_proxy) { in is_qp0() 623 qp->mqp.usage = MLX4_RES_USAGE_USER_VERBS; in create_qp_rss() [all …]
|
D | mlx4_ib.h | 327 struct mlx4_qp mqp; member 718 static inline struct mlx4_ib_qp *to_mibqp(struct mlx4_qp *mqp) in to_mibqp() argument 720 return container_of(mqp, struct mlx4_ib_qp, mqp); in to_mibqp() 850 int mlx4_ib_add_mc(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp, 908 int mlx4_ib_steer_qp_reg(struct mlx4_ib_dev *mdev, struct mlx4_ib_qp *mqp,
|
D | cq.c | 661 struct mlx4_qp *mqp; in mlx4_ib_poll_one() local 709 (be32_to_cpu(cqe->vlan_my_qpn) & MLX4_CQE_QPN_MASK) != (*cur_qp)->mqp.qpn) { in mlx4_ib_poll_one() 715 mqp = __mlx4_qp_lookup(to_mdev(cq->ibcq.device)->dev, in mlx4_ib_poll_one() 717 *cur_qp = to_mibqp(mqp); in mlx4_ib_poll_one()
|
/Linux-v5.15/arch/sparc/kernel/ |
D | pci_msi.c | 335 } *mqp; in sparc64_pbm_msi_init() local 355 mqp = of_get_property(pbm->op->dev.of_node, in sparc64_pbm_msi_init() 357 if (!mqp) in sparc64_pbm_msi_init() 358 mqp = of_get_property(pbm->op->dev.of_node, in sparc64_pbm_msi_init() 360 if (!mqp || len != sizeof(struct msiq_prop)) in sparc64_pbm_msi_init() 363 pbm->msiq_first = mqp->first_msiq; in sparc64_pbm_msi_init() 364 pbm->msiq_first_devino = mqp->first_devino; in sparc64_pbm_msi_init()
|
/Linux-v5.15/drivers/vdpa/mlx5/net/ |
D | mlx5_vnet.c | 80 struct mlx5_core_qp mqp; member 474 vqp->mqp.uid = ndev->mvdev.res.uid; in qp_create() 475 vqp->mqp.qpn = MLX5_GET(create_qp_out, out, qpn); in qp_create() 497 MLX5_SET(destroy_qp_in, in, qpn, vqp->mqp.qpn); in qp_destroy() 500 mlx5_vdpa_warn(&ndev->mvdev, "destroy qp 0x%x\n", vqp->mqp.qpn); in qp_destroy() 861 MLX5_SET(virtio_q, vq_ctx, event_qpn_or_msix, mvq->fwqp.mqp.qpn); in create_virtqueue() 915 return fw ? mvq->vqqp.mqp.qpn : mvq->fwqp.mqp.qpn; in get_rqpn() 920 return fw ? mvq->fwqp.mqp.qpn : mvq->vqqp.mqp.qpn; in get_qpn()
|
/Linux-v5.15/include/linux/mlx5/ |
D | qp.h | 498 struct mlx5_core_qp mqp; member
|