/Linux-v5.4/drivers/net/ethernet/mellanox/mlx4/ |
D | cq.c | 104 void mlx4_cq_completion(struct mlx4_dev *dev, u32 cqn) in mlx4_cq_completion() argument 110 cqn & (dev->caps.num_cqs - 1)); in mlx4_cq_completion() 114 mlx4_dbg(dev, "Completion event for bogus CQ %08x\n", cqn); in mlx4_cq_completion() 126 void mlx4_cq_event(struct mlx4_dev *dev, u32 cqn, int event_type) in mlx4_cq_event() argument 132 cq = radix_tree_lookup(&cq_table->tree, cqn & (dev->caps.num_cqs - 1)); in mlx4_cq_event() 136 mlx4_dbg(dev, "Async event for bogus CQ %08x\n", cqn); in mlx4_cq_event() 184 err = mlx4_MODIFY_CQ(dev, mailbox, cq->cqn, 1); in mlx4_cq_modify() 210 err = mlx4_MODIFY_CQ(dev, mailbox, cq->cqn, 0); in mlx4_cq_resize() 217 int __mlx4_cq_alloc_icm(struct mlx4_dev *dev, int *cqn) in __mlx4_cq_alloc_icm() argument 223 *cqn = mlx4_bitmap_alloc(&cq_table->bitmap); in __mlx4_cq_alloc_icm() [all …]
|
D | en_resources.c | 41 int is_tx, int rss, int qpn, int cqn, in mlx4_en_fill_qp_context() argument 72 context->cqn_send = cpu_to_be32(cqn); in mlx4_en_fill_qp_context() 73 context->cqn_recv = cpu_to_be32(cqn); in mlx4_en_fill_qp_context()
|
D | resource_tracker.c | 1671 static int cq_res_start_move_to(struct mlx4_dev *dev, int slave, int cqn, in cq_res_start_move_to() argument 1680 r = res_tracker_lookup(&tracker->res_tree[RES_CQ], cqn); in cq_res_start_move_to() 1938 int cqn; in cq_alloc_res() local 1947 err = __mlx4_cq_alloc_icm(dev, &cqn); in cq_alloc_res() 1953 err = add_res_range(dev, slave, cqn, 1, RES_CQ, 0); in cq_alloc_res() 1956 __mlx4_cq_free_icm(dev, cqn); in cq_alloc_res() 1960 set_param_l(out_param, cqn); in cq_alloc_res() 2464 int cqn; in cq_free_res() local 2469 cqn = get_param_l(&in_param); in cq_free_res() 2470 err = rem_res_range(dev, slave, cqn, 1, RES_CQ, 0); in cq_free_res() [all …]
|
D | srq.c | 162 int mlx4_srq_alloc(struct mlx4_dev *dev, u32 pdn, u32 cqn, u16 xrcd, in mlx4_srq_alloc() argument 192 srq_context->pg_offset_cqn = cpu_to_be32(cqn & 0xffffff); in mlx4_srq_alloc()
|
/Linux-v5.4/drivers/infiniband/hw/hns/ |
D | hns_roce_cq.c | 64 event_type, hr_cq->cqn); in hns_roce_ib_cq_event() 117 ret = hns_roce_bitmap_alloc(&cq_table->bitmap, &hr_cq->cqn); in hns_roce_cq_alloc() 124 ret = hns_roce_table_get(hr_dev, &cq_table->table, hr_cq->cqn); in hns_roce_cq_alloc() 130 ret = xa_err(xa_store(&cq_table->array, hr_cq->cqn, hr_cq, GFP_KERNEL)); in hns_roce_cq_alloc() 147 ret = hns_roce_sw2hw_cq(hr_dev, mailbox, hr_cq->cqn); in hns_roce_cq_alloc() 163 xa_erase(&cq_table->array, hr_cq->cqn); in hns_roce_cq_alloc() 166 hns_roce_table_put(hr_dev, &cq_table->table, hr_cq->cqn); in hns_roce_cq_alloc() 169 hns_roce_bitmap_free(&cq_table->bitmap, hr_cq->cqn, BITMAP_NO_RR); in hns_roce_cq_alloc() 188 ret = hns_roce_hw2sw_cq(hr_dev, NULL, hr_cq->cqn); in hns_roce_free_cq() 191 hr_cq->cqn); in hns_roce_free_cq() [all …]
|
D | hns_roce_hw_v2_dfx.c | 9 int hns_roce_v2_query_cqc_info(struct hns_roce_dev *hr_dev, u32 cqn, in hns_roce_v2_query_cqc_info() argument 21 ret = hns_roce_cmd_mbox(hr_dev, 0, mailbox->dma, cqn, 0, in hns_roce_v2_query_cqc_info()
|
D | hns_roce_srq.c | 80 static int hns_roce_srq_alloc(struct hns_roce_dev *hr_dev, u32 pdn, u32 cqn, in hns_roce_srq_alloc() argument 133 hr_dev->hw->write_srqc(hr_dev, srq, pdn, xrcd, cqn, mailbox->buf, in hns_roce_srq_alloc() 366 u32 cqn; in hns_roce_create_srq() local 404 cqn = ib_srq_has_cq(srq_init_attr->srq_type) ? in hns_roce_create_srq() 405 to_hr_cq(srq_init_attr->ext.cq)->cqn : 0; in hns_roce_create_srq() 409 ret = hns_roce_srq_alloc(hr_dev, to_hr_pd(ib_srq->pd)->pdn, cqn, 0, in hns_roce_create_srq()
|
D | hns_roce_device.h | 507 unsigned long cqn; member 919 u32 cqn; member 925 int (*query_cqc_info)(struct hns_roce_dev *hr_dev, u32 cqn, 983 struct hns_roce_srq *srq, u32 pdn, u16 xrcd, u32 cqn, 1276 void hns_roce_cq_completion(struct hns_roce_dev *hr_dev, u32 cqn); 1277 void hns_roce_cq_event(struct hns_roce_dev *hr_dev, u32 cqn, int event_type);
|
D | hns_roce_restrack.c | 96 ret = hr_dev->dfx->query_cqc_info(hr_dev, hr_cq->cqn, (int *)context); in hns_roce_fill_res_cq_entry()
|
D | hns_roce_hw_v1.c | 2012 ROCEE_DB_OTHERS_H_ROCEE_DB_OTH_INP_H_S, hr_cq->cqn); in hns_roce_v1_cq_set_ci() 2091 offset = hr_cq->cqn * HNS_ROCE_V1_TPTR_ENTRY_SIZE; in hns_roce_v1_write_cqc() 2100 CQ_CONTEXT_CQC_BYTE_4_CQN_S, hr_cq->cqn); in hns_roce_v1_write_cqc() 2181 hr_cq->cqn | notification_flag); in hns_roce_v1_req_notify_cq() 2231 hr_cq->cqn, (qpn & HNS_ROCE_CQE_QPN_MASK)); in hns_roce_v1_poll_one() 2650 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_sqp() 2654 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_sqp() 2800 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_qp() 2804 to_hr_cq(ibqp->recv_cq)->cqn); in hns_roce_v1_m_qp() 2866 to_hr_cq(ibqp->send_cq)->cqn); in hns_roce_v1_m_qp() [all …]
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/ |
D | cq.c | 109 cq->cqn = MLX5_GET(create_cq_out, out, cqn); in mlx5_core_create_cq() 136 cq->cqn); in mlx5_core_create_cq() 148 MLX5_SET(destroy_cq_in, din, cqn, cq->cqn); in mlx5_core_create_cq() 165 MLX5_SET(destroy_cq_in, in, cqn, cq->cqn); in mlx5_core_destroy_cq() 187 MLX5_SET(query_cq_in, in, cqn, cq->cqn); in mlx5_core_query_cq() 211 MLX5_SET(modify_cq_in, in, cqn, cq->cqn); in mlx5_core_modify_cq_moderation()
|
D | eq.c | 114 static struct mlx5_core_cq *mlx5_eq_cq_get(struct mlx5_eq *eq, u32 cqn) in mlx5_eq_cq_get() argument 120 cq = radix_tree_lookup(&table->tree, cqn); in mlx5_eq_cq_get() 137 u32 cqn = -1; in mlx5_eq_comp_int() local 151 cqn = be32_to_cpu(eqe->data.comp.cqn) & 0xffffff; in mlx5_eq_comp_int() 153 cq = mlx5_eq_cq_get(eq, cqn); in mlx5_eq_comp_int() 159 mlx5_core_warn(eq->dev, "Completion event for bogus CQ 0x%x\n", cqn); in mlx5_eq_comp_int() 169 if (cqn != -1) in mlx5_eq_comp_int() 389 err = radix_tree_insert(&table->tree, cq->cqn, cq); in mlx5_eq_add_cq() 401 tmp = radix_tree_delete(&table->tree, cq->cqn); in mlx5_eq_del_cq() 406 eq->eqn, cq->cqn); in mlx5_eq_del_cq() [all …]
|
/Linux-v5.4/drivers/infiniband/hw/mthca/ |
D | mthca_cq.c | 76 __be32 cqn; member 211 mthca_write64(MTHCA_TAVOR_CQ_DB_INC_CI | cq->cqn, incr - 1, in update_cons_index() 217 void mthca_cq_completion(struct mthca_dev *dev, u32 cqn) in mthca_cq_completion() argument 221 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); in mthca_cq_completion() 224 mthca_warn(dev, "Completion event for bogus CQ %08x\n", cqn); in mthca_cq_completion() 233 void mthca_cq_event(struct mthca_dev *dev, u32 cqn, in mthca_cq_event() argument 241 cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1)); in mthca_cq_event() 248 mthca_warn(dev, "Async event for bogus CQ %08x\n", cqn); in mthca_cq_event() 297 qpn, cq->cqn, cq->cons_index, prod_index); in mthca_cq_clean() 384 cq->cqn, cq->cons_index); in handle_error_cqe() [all …]
|
D | mthca_eq.c | 132 __be32 cqn; member 149 __be32 cqn; member 219 static inline void disarm_cq(struct mthca_dev *dev, int eqn, int cqn) in disarm_cq() argument 222 mthca_write64(MTHCA_EQ_DB_DISARM_CQ | eqn, cqn, in disarm_cq() 276 disarm_cqn = be32_to_cpu(eqe->event.comp.cqn) & 0xffffff; in mthca_eq_int() 343 be32_to_cpu(eqe->event.cq_err.cqn) & 0xffffff); in mthca_eq_int() 344 mthca_cq_event(dev, be32_to_cpu(eqe->event.cq_err.cqn), in mthca_eq_int()
|
/Linux-v5.4/drivers/infiniband/hw/mlx4/ |
D | srq.c | 82 u32 cqn; in mlx4_ib_create_srq() local 174 cqn = ib_srq_has_cq(init_attr->srq_type) ? in mlx4_ib_create_srq() 175 to_mcq(init_attr->ext.cq)->mcq.cqn : 0; in mlx4_ib_create_srq() 179 err = mlx4_srq_alloc(dev->dev, to_mpd(ib_srq->pd)->pdn, cqn, xrcdn, in mlx4_ib_create_srq()
|
/Linux-v5.4/include/linux/mlx5/ |
D | cq.h | 41 u32 cqn; member 171 doorbell[1] = cpu_to_be32(cq->cqn); in mlx5_cq_arm()
|
/Linux-v5.4/include/uapi/rdma/ |
D | hns-abi.h | 45 __aligned_u64 cqn; /* Only 32 bits used, 64 for compat */ member
|
D | mthca-abi.h | 84 __u32 cqn; member
|
D | mlx4-abi.h | 85 __u32 cqn; member
|
/Linux-v5.4/drivers/infiniband/hw/mlx5/ |
D | cq.c | 56 type, mcq->cqn); in mlx5_ib_cq_event() 498 "Requestor" : "Responder", cq->mcq.cqn); in mlx5_poll_one() 534 cq->mcq.cqn, mr->sig->err_item.key, in mlx5_poll_one() 559 cq->mcq.cqn); in poll_soft_wc() 962 mlx5_ib_dbg(dev, "cqn 0x%x\n", cq->mcq.cqn); in mlx5_ib_create_cq() 973 if (ib_copy_to_udata(udata, &cq->mcq.cqn, sizeof(__u32))) { in mlx5_ib_create_cq() 1087 mlx5_ib_warn(dev, "modify cq 0x%x failed\n", mcq->mcq.cqn); in mlx5_ib_modify_cq() 1198 cq->mcq.cqn); in copy_resize_cqes() 1291 MLX5_SET(modify_cq_in, in, cqn, cq->mcq.cqn); in mlx5_ib_resize_cq()
|
D | srq.c | 295 in.cqn = to_mcq(init_attr->ext.cq)->mcq.cqn; in mlx5_ib_create_srq() 297 in.cqn = to_mcq(dev->devr.c0)->mcq.cqn; in mlx5_ib_create_srq()
|
D | srq.h | 25 u32 cqn; member
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | dr_send.c | 43 u32 cqn; member 169 MLX5_SET(qpc, qpc, cqn_snd, attr->cqn); in dr_create_rc_qp() 170 MLX5_SET(qpc, qpc, cqn_rcv, attr->cqn); in dr_create_rc_qp() 687 pr_info("CQ event %u on CQ #%u\n", event, mcq->cqn); in dr_cq_event() 867 init_attr.cqn = dmn->send_ring->cq->mcq.cqn; in mlx5dr_send_ring_alloc()
|
/Linux-v5.4/include/linux/mlx4/ |
D | cq.h | 161 doorbell[0] = cpu_to_be32(sn << 28 | cmd | cq->cqn); in mlx4_cq_arm()
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
D | conn.c | 371 mlx5_fpga_warn(conn->fdev, "CQ event %u on CQ #%u\n", event, mcq->cqn); in mlx5_fpga_conn_cq_event() 502 mlx5_fpga_dbg(fdev, "Created CQ #0x%x\n", conn->cq.mcq.cqn); in mlx5_fpga_conn_create_cq() 594 MLX5_SET(qpc, qpc, cqn_snd, conn->cq.mcq.cqn); in mlx5_fpga_conn_create_qp() 595 MLX5_SET(qpc, qpc, cqn_rcv, conn->cq.mcq.cqn); in mlx5_fpga_conn_create_qp() 699 MLX5_SET(qpc, qpc, cqn_snd, conn->cq.mcq.cqn); in mlx5_fpga_conn_init_qp() 700 MLX5_SET(qpc, qpc, cqn_rcv, conn->cq.mcq.cqn); in mlx5_fpga_conn_init_qp()
|