/Linux-v4.19/drivers/net/ethernet/mellanox/mlx4/ |
D | qp.c | 49 void mlx4_qp_event(struct mlx4_dev *dev, u32 qpn, int event_type) in mlx4_qp_event() argument 56 qp = __mlx4_qp_lookup(dev, qpn); in mlx4_qp_event() 63 mlx4_dbg(dev, "Async event for none existent QP %08x\n", qpn); in mlx4_qp_event() 79 *proxy_qp0 = qp->qpn >= pf_proxy_offset && qp->qpn <= pf_proxy_offset + 1; in is_master_qp0() 81 *real_qp0 = qp->qpn >= dev->phys_caps.base_sqpn && in is_master_qp0() 82 qp->qpn <= dev->phys_caps.base_sqpn + 1; in is_master_qp0() 145 ret = mlx4_cmd(dev, 0, qp->qpn, 2, in __mlx4_qp_modify() 150 port = (qp->qpn & 1) + 1; in __mlx4_qp_modify() 174 cpu_to_be16(mlx4_qp_roce_entropy(dev, qp->qpn)); in __mlx4_qp_modify() 180 cpu_to_be32(qp->qpn); in __mlx4_qp_modify() [all …]
|
D | mcg.c | 125 u32 qpn) in get_promisc_qp() argument 136 if (pqp->qpn == qpn) in get_promisc_qp() 149 unsigned int index, u32 qpn) in new_steering_entry() argument 176 pqp = get_promisc_qp(dev, port, steer, qpn); in new_steering_entry() 183 dqp->qpn = qpn; in new_steering_entry() 209 if (pqp->qpn == qpn) in new_steering_entry() 218 mgm->qp[members_count++] = cpu_to_be32(pqp->qpn & MGM_QPN_MASK); in new_steering_entry() 241 unsigned int index, u32 qpn) in existing_steering_entry() argument 253 pqp = get_promisc_qp(dev, port, steer, qpn); in existing_steering_entry() 272 if (qpn == dqp->qpn) in existing_steering_entry() [all …]
|
D | en_resources.c | 41 int is_tx, int rss, int qpn, int cqn, in mlx4_en_fill_qp_context() argument 63 context->local_qpn = cpu_to_be32(qpn); in mlx4_en_fill_qp_context() 89 en_dbg(HW, priv, "Setting RX qp %x tunnel mode to RX tunneled & non-tunneled\n", qpn); in mlx4_en_fill_qp_context() 104 ret = mlx4_update_qp(priv->mdev->dev, qp->qpn, in mlx4_en_change_mcast_lb()
|
D | resource_tracker.c | 225 int qpn; member 741 u8 slave, u32 qpn) in update_vport_qp_param() argument 762 if (mlx4_is_qp_reserved(dev, qpn)) in update_vport_qp_param() 776 err = mlx4_update_qp(dev, qpn, MLX4_UPDATE_QP_VSD, ¶ms); in update_vport_qp_param() 1162 static struct res_common *alloc_fs_rule_tr(u64 id, int qpn) in alloc_fs_rule_tr() argument 1172 ret->qpn = qpn; in alloc_fs_rule_tr() 1494 static int qp_res_start_move_to(struct mlx4_dev *dev, int slave, int qpn, in qp_res_start_move_to() argument 1504 r = res_tracker_lookup(&tracker->res_tree[RES_QP], qpn); in qp_res_start_move_to() 1765 static int valid_reserved(struct mlx4_dev *dev, int slave, int qpn) in valid_reserved() argument 1767 return mlx4_is_qp_reserved(dev, qpn) && in valid_reserved() [all …]
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/ |
D | qp.c | 167 qp->qpn | (rsc_type << MLX5_USER_INDEX_LEN), in create_resource_common() 188 qp->qpn | (qp->common.res << MLX5_USER_INDEX_LEN)); in destroy_resource_common() 213 qp->qpn = MLX5_GET(create_dct_out, out, dctn); in mlx5_core_create_dct() 221 MLX5_SET(destroy_dct_in, din, dctn, qp->qpn); in mlx5_core_create_dct() 243 qp->qpn = MLX5_GET(create_qp_out, out, qpn); in mlx5_core_create_qp() 244 mlx5_core_dbg(dev, "qpn = 0x%x\n", qp->qpn); in mlx5_core_create_qp() 253 qp->qpn); in mlx5_core_create_qp() 263 MLX5_SET(destroy_qp_in, din, qpn, qp->qpn); in mlx5_core_create_qp() 277 MLX5_SET(drain_dct_in, in, dctn, qp->qpn); in mlx5_core_drain_dct() 295 mlx5_core_warn(dev, "failed drain DCT 0x%x with error 0x%x\n", qp->qpn, err); in mlx5_core_destroy_dct() [all …]
|
D | mcg.c | 40 int mlx5_core_attach_mcg(struct mlx5_core_dev *dev, union ib_gid *mgid, u32 qpn) in mlx5_core_attach_mcg() argument 47 MLX5_SET(attach_to_mcg_in, in, qpn, qpn); in mlx5_core_attach_mcg() 54 int mlx5_core_detach_mcg(struct mlx5_core_dev *dev, union ib_gid *mgid, u32 qpn) in mlx5_core_detach_mcg() argument 61 MLX5_SET(detach_from_mcg_in, in, qpn, qpn); in mlx5_core_detach_mcg()
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/ipoib/ |
D | ipoib_vlan.c | 72 u32 qpn) in mlx5i_find_qpn_to_netdev_node() argument 74 struct hlist_head *h = &buckets[hash_32(qpn, MLX5I_MAX_LOG_PKEY_SUP)]; in mlx5i_find_qpn_to_netdev_node() 78 if (node->underlay_qpn == qpn) in mlx5i_find_qpn_to_netdev_node() 85 int mlx5i_pkey_add_qpn(struct net_device *netdev, u32 qpn) in mlx5i_pkey_add_qpn() argument 89 u8 key = hash_32(qpn, MLX5I_MAX_LOG_PKEY_SUP); in mlx5i_pkey_add_qpn() 97 new_node->underlay_qpn = qpn; in mlx5i_pkey_add_qpn() 105 int mlx5i_pkey_del_qpn(struct net_device *netdev, u32 qpn) in mlx5i_pkey_del_qpn() argument 112 node = mlx5i_find_qpn_to_netdev_node(ht->buckets, qpn); in mlx5i_pkey_del_qpn() 126 struct net_device *mlx5i_pkey_get_netdev(struct net_device *netdev, u32 qpn) in mlx5i_pkey_get_netdev() argument 131 node = mlx5i_find_qpn_to_netdev_node(ipriv->qpn_htbl->buckets, qpn); in mlx5i_pkey_get_netdev() [all …]
|
D | ipoib.c | 228 err = mlx5e_create_tis(priv->mdev, 0 /* tc */, ipriv->qp.qpn, &priv->tisn[0]); in mlx5i_init_tx() 409 dev->dev_addr[1] = (ipriv->qp.qpn >> 16) & 0xff; in mlx5i_dev_init() 410 dev->dev_addr[2] = (ipriv->qp.qpn >> 8) & 0xff; in mlx5i_dev_init() 411 dev->dev_addr[3] = (ipriv->qp.qpn) & 0xff; in mlx5i_dev_init() 414 mlx5i_pkey_add_qpn(dev ,ipriv->qp.qpn); in mlx5i_dev_init() 441 mlx5i_pkey_del_qpn(dev, ipriv->qp.qpn); in mlx5i_dev_cleanup() 461 err = mlx5_fs_add_rx_underlay_qpn(mdev, ipriv->qp.qpn); in mlx5i_open() 478 mlx5_fs_remove_rx_underlay_qpn(mdev, ipriv->qp.qpn); in mlx5i_open() 504 mlx5_fs_remove_rx_underlay_qpn(mdev, ipriv->qp.qpn); in mlx5i_close() 523 mlx5_core_dbg(mdev, "attaching QPN 0x%x, MGID %pI6\n", ipriv->qp.qpn, gid->raw); in mlx5i_attach_mcast() [all …]
|
D | ipoib.h | 75 int mlx5i_pkey_add_qpn(struct net_device *netdev, u32 qpn); 76 int mlx5i_pkey_del_qpn(struct net_device *netdev, u32 qpn); 79 struct net_device *mlx5i_pkey_get_netdev(struct net_device *netdev, u32 qpn);
|
/Linux-v4.19/drivers/infiniband/hw/qib/ |
D | qib_qp.c | 130 u32 i, offset, max_scan, qpn; in qib_alloc_qpn() local 152 qpn = qpt->last + 2; in qib_alloc_qpn() 153 if (qpn >= RVT_QPN_MAX) in qib_alloc_qpn() 154 qpn = 2; in qib_alloc_qpn() 155 if (qpt_mask && ((qpn & qpt_mask) >> 1) >= dd->n_krcv_queues) in qib_alloc_qpn() 156 qpn = (qpn | qpt_mask) + 2; in qib_alloc_qpn() 157 offset = qpn & RVT_BITS_PER_PAGE_MASK; in qib_alloc_qpn() 158 map = &qpt->map[qpn / RVT_BITS_PER_PAGE]; in qib_alloc_qpn() 168 qpt->last = qpn; in qib_alloc_qpn() 169 ret = qpn; in qib_alloc_qpn() [all …]
|
/Linux-v4.19/drivers/infiniband/sw/rdmavt/ |
D | trace_qp.h | 64 __field(u32, qpn) 69 __entry->qpn = qp->ibqp.qp_num; 75 __entry->qpn, 94 __field(u32, qpn) 101 __entry->qpn = qp->ibqp.qp_num; 109 __entry->qpn,
|
D | trace_rc.h | 64 __field(u32, qpn) 75 __entry->qpn = qp->ibqp.qp_num; 87 __entry->qpn,
|
/Linux-v4.19/drivers/infiniband/sw/rxe/ |
D | rxe_recv.c | 102 u32 qpn, struct rxe_qp *qp) in check_keys() argument 111 if (qpn == 1) { in check_keys() 125 } else if (qpn != 0) { in check_keys() 137 qpn != 0 && pkt->mask) { in check_keys() 138 u32 qkey = (qpn == 1) ? GSI_QKEY : qp->attr.qkey; in check_keys() 142 deth_qkey(pkt), qkey, qpn); in check_keys() 219 u32 qpn = bth_qpn(pkt); in hdr_check() local 228 if (unlikely(qpn == 0)) { in hdr_check() 233 if (qpn != IB_MULTICAST_QPN) { in hdr_check() 234 index = (qpn == 1) ? port->qp_gsi_index : qpn; in hdr_check() [all …]
|
D | rxe_hdr.h | 93 __be32 qpn; member 208 return BTH_QPN_MASK & be32_to_cpu(bth->qpn); in __bth_qpn() 211 static inline void __bth_set_qpn(void *arg, u32 qpn) in __bth_set_qpn() argument 214 u32 resvqpn = be32_to_cpu(bth->qpn); in __bth_set_qpn() 216 bth->qpn = cpu_to_be32((BTH_QPN_MASK & qpn) | in __bth_set_qpn() 224 return 0 != (cpu_to_be32(BTH_FECN_MASK) & bth->qpn); in __bth_fecn() 232 bth->qpn |= cpu_to_be32(BTH_FECN_MASK); in __bth_set_fecn() 234 bth->qpn &= ~cpu_to_be32(BTH_FECN_MASK); in __bth_set_fecn() 241 return 0 != (cpu_to_be32(BTH_BECN_MASK) & bth->qpn); in __bth_becn() 249 bth->qpn |= cpu_to_be32(BTH_BECN_MASK); in __bth_set_becn() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/hns/ |
D | hns_roce_qp.c | 44 void hns_roce_qp_event(struct hns_roce_dev *hr_dev, u32 qpn, int event_type) in hns_roce_qp_event() argument 52 qp = __hns_roce_qp_lookup(hr_dev, qpn); in hns_roce_qp_event() 59 dev_warn(dev, "Async event for bogus QP %08x\n", qpn); in hns_roce_qp_event() 106 type, hr_qp->qpn); in hns_roce_ib_qp_event() 145 static int hns_roce_gsi_qp_alloc(struct hns_roce_dev *hr_dev, unsigned long qpn, in hns_roce_gsi_qp_alloc() argument 151 if (!qpn) in hns_roce_gsi_qp_alloc() 154 hr_qp->qpn = qpn; in hns_roce_gsi_qp_alloc() 158 hr_qp->qpn & (hr_dev->caps.num_qps - 1), hr_qp); in hns_roce_gsi_qp_alloc() 175 static int hns_roce_qp_alloc(struct hns_roce_dev *hr_dev, unsigned long qpn, in hns_roce_qp_alloc() argument 182 if (!qpn) in hns_roce_qp_alloc() [all …]
|
D | hns_roce_hw_v1.c | 434 RQ_DOORBELL_U32_8_QPN_S, hr_qp->qpn); in hns_roce_v1_post_recv() 813 attr.dest_qp_num = hr_qp->qpn; in hns_roce_v1_rsv_lp_qp() 1059 hr_qp->qpn, ret); in hns_roce_v1_mr_free_work_fn() 1074 hr_qp->qpn, ret, hr_mr->key, ne); in hns_roce_v1_mr_free_work_fn() 2050 static void __hns_roce_v1_cq_clean(struct hns_roce_cq *hr_cq, u32 qpn, in __hns_roce_v1_cq_clean() argument 2072 HNS_ROCE_CQE_QPN_MASK) == qpn) { in __hns_roce_v1_cq_clean() 2098 static void hns_roce_v1_cq_clean(struct hns_roce_cq *hr_cq, u32 qpn, in hns_roce_v1_cq_clean() argument 2102 __hns_roce_v1_cq_clean(hr_cq, qpn, srq); in hns_roce_v1_cq_clean() 2224 int qpn; in hns_roce_v1_poll_one() local 2250 qpn = roce_get_field(cqe->cqe_byte_20, CQE_BYTE_20_PORT_NUM_M, in hns_roce_v1_poll_one() [all …]
|
/Linux-v4.19/include/uapi/rdma/ |
D | ib_user_mad.h | 79 __be32 qpn; member 123 __be32 qpn; member 188 __u8 qpn; member 223 __u32 qpn; member
|
D | ib_user_cm.h | 134 __u32 qpn; member 159 __u32 qpn; member 212 __u32 qpn; member 296 __u32 qpn; member
|
/Linux-v4.19/drivers/infiniband/hw/mthca/ |
D | mthca_qp.c | 197 return qp->qpn >= dev->qp_table.sqp_start && in is_sqp() 198 qp->qpn <= dev->qp_table.sqp_start + 3; in is_sqp() 203 return qp->qpn >= dev->qp_table.sqp_start && in is_qp0() 204 qp->qpn <= dev->qp_table.sqp_start + 1; in is_qp0() 237 void mthca_qp_event(struct mthca_dev *dev, u32 qpn, in mthca_qp_event() argument 244 qp = mthca_array_get(&dev->qp_table.qp, qpn & (dev->limits.num_qps - 1)); in mthca_qp_event() 251 event_type, qpn); in mthca_qp_event() 453 err = mthca_QUERY_QP(dev, qp->qpn, 0, mailbox); in mthca_query_qp() 626 qp_context->local_qpn = cpu_to_be32(qp->qpn); in __mthca_modify_qp() 767 ((qp->qpn & (dev->limits.num_qps - 1)) * MTHCA_RDB_ENTRY_SIZE << in __mthca_modify_qp() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/hfi1/ |
D | trace_ibhdrs.h | 108 u16 *pkey, u32 *psn, u32 *qpn); 115 u32 *psn, u32 *qpn); 131 u8 tver, u16 pkey, u32 psn, u32 qpn, 174 __field(u32, qpn) 216 &__entry->qpn); 240 &__entry->qpn); 287 __entry->qpn, 333 __field(u32, qpn) 381 &__entry->qpn); 408 &__entry->qpn); [all …]
|
D | trace_rc.h | 63 __field(u32, qpn) 74 __entry->qpn = qp->ibqp.qp_num; 86 __entry->qpn,
|
D | trace.c | 150 u16 *pkey, u32 *psn, u32 *qpn) in hfi1_trace_parse_9b_bth() argument 162 *qpn = ib_bth_get_qpn(ohdr); in hfi1_trace_parse_9b_bth() 168 u32 *psn, u32 *qpn) in hfi1_trace_parse_16b_bth() argument 177 *qpn = ib_bth_get_qpn(ohdr); in hfi1_trace_parse_16b_bth() 249 u8 tver, u16 pkey, u32 psn, u32 qpn, in hfi1_trace_fmt_rest() argument 261 se, mig, pad, tver, qpn, ack, psn); in hfi1_trace_fmt_rest() 267 qpn, ack, psn); in hfi1_trace_fmt_rest()
|
/Linux-v4.19/include/rdma/ |
D | iw_cm.h | 105 u32 qpn; member 114 int qpn); 172 struct ib_qp *iw_cm_get_qp(struct ib_device *device, int qpn);
|
/Linux-v4.19/drivers/infiniband/core/ |
D | cm_msgs.h | 108 static inline void cm_req_set_local_qpn(struct cm_req_msg *req_msg, __be32 qpn) in cm_req_set_local_qpn() argument 110 req_msg->offset32 = cpu_to_be32((be32_to_cpu(qpn) << 8) | in cm_req_set_local_qpn() 516 static inline void cm_rep_set_local_qpn(struct cm_rep_msg *rep_msg, __be32 qpn) in cm_rep_set_local_qpn() argument 518 rep_msg->offset12 = cpu_to_be32((be32_to_cpu(qpn) << 8) | in cm_rep_set_local_qpn() 636 static inline void cm_dreq_set_remote_qpn(struct cm_dreq_msg *dreq_msg, __be32 qpn) in cm_dreq_set_remote_qpn() argument 638 dreq_msg->offset8 = cpu_to_be32((be32_to_cpu(qpn) << 8) | in cm_dreq_set_remote_qpn() 685 static inline void cm_lap_set_remote_qpn(struct cm_lap_msg *lap_msg, __be32 qpn) in cm_lap_set_remote_qpn() argument 687 lap_msg->offset12 = cpu_to_be32((be32_to_cpu(qpn) << 8) | in cm_lap_set_remote_qpn() 822 __be32 qpn) in cm_sidr_rep_set_qpn() argument 824 sidr_rep_msg->offset8 = cpu_to_be32((be32_to_cpu(qpn) << 8) | in cm_sidr_rep_set_qpn()
|
/Linux-v4.19/include/linux/mlx4/ |
D | device.h | 767 int qpn; member 928 __be32 qpn; member 1056 static inline int mlx4_is_qp_reserved(struct mlx4_dev *dev, u32 qpn) in mlx4_is_qp_reserved() argument 1058 return (qpn < dev->phys_caps.base_sqpn + 8 + in mlx4_is_qp_reserved() 1060 qpn >= dev->phys_caps.base_sqpn) || in mlx4_is_qp_reserved() 1061 (qpn < dev->caps.reserved_qps_cnt[MLX4_QP_REGION_FW]); in mlx4_is_qp_reserved() 1064 static inline int mlx4_is_guest_proxy(struct mlx4_dev *dev, int slave, u32 qpn) in mlx4_is_guest_proxy() argument 1068 if (qpn >= guest_proxy_base && qpn < guest_proxy_base + 8) in mlx4_is_guest_proxy() 1145 int mlx4_qp_alloc(struct mlx4_dev *dev, int qpn, struct mlx4_qp *qp); 1272 u32 qpn; member [all …]
|