/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/ |
D | transobj.c | 116 int mlx5_core_create_sq(struct mlx5_core_dev *dev, u32 *in, int inlen, u32 *sqn) in mlx5_core_create_sq() argument 124 *sqn = MLX5_GET(create_sq_out, out, sqn); in mlx5_core_create_sq() 129 int mlx5_core_modify_sq(struct mlx5_core_dev *dev, u32 sqn, u32 *in, int inlen) in mlx5_core_modify_sq() argument 133 MLX5_SET(modify_sq_in, in, sqn, sqn); in mlx5_core_modify_sq() 139 void mlx5_core_destroy_sq(struct mlx5_core_dev *dev, u32 sqn) in mlx5_core_destroy_sq() argument 145 MLX5_SET(destroy_sq_in, in, sqn, sqn); in mlx5_core_destroy_sq() 149 int mlx5_core_query_sq(struct mlx5_core_dev *dev, u32 sqn, u32 *out) in mlx5_core_query_sq() argument 155 MLX5_SET(query_sq_in, in, sqn, sqn); in mlx5_core_query_sq() 160 int mlx5_core_query_sq_state(struct mlx5_core_dev *dev, u32 sqn, u8 *state) in mlx5_core_query_sq_state() argument 172 err = mlx5_core_query_sq(dev, sqn, out); in mlx5_core_query_sq_state() [all …]
|
D | qp.c | 639 static void destroy_sq_tracked(struct mlx5_core_dev *dev, u32 sqn, u16 uid) in destroy_sq_tracked() argument 645 MLX5_SET(destroy_sq_in, in, sqn, sqn); in destroy_sq_tracked() 654 u32 sqn; in mlx5_core_create_sq_tracked() local 656 err = mlx5_core_create_sq(dev, in, inlen, &sqn); in mlx5_core_create_sq_tracked() 661 sq->qpn = sqn; in mlx5_core_create_sq_tracked()
|
D | en_main.c | 240 cseg->qpn_ds = cpu_to_be32((sq->sqn << MLX5_WQE_CTRL_QPN_SHIFT) | in mlx5e_build_umr_wqe() 1186 u32 *sqn) in mlx5e_create_sq() argument 1223 err = mlx5_core_create_sq(mdev, in, inlen, sqn); in mlx5e_create_sq() 1230 int mlx5e_modify_sq(struct mlx5_core_dev *mdev, u32 sqn, in mlx5e_modify_sq() argument 1252 err = mlx5_core_modify_sq(mdev, sqn, in, inlen); in mlx5e_modify_sq() 1259 static void mlx5e_destroy_sq(struct mlx5_core_dev *mdev, u32 sqn) in mlx5e_destroy_sq() argument 1261 mlx5_core_destroy_sq(mdev, sqn); in mlx5e_destroy_sq() 1267 u32 *sqn) in mlx5e_create_sq_rdy() argument 1272 err = mlx5e_create_sq(mdev, param, csp, sqn); in mlx5e_create_sq_rdy() 1278 err = mlx5e_modify_sq(mdev, *sqn, &msp); in mlx5e_create_sq_rdy() [all …]
|
D | en.h | 407 u32 sqn; member 531 u32 sqn; member 559 u32 sqn; member 1064 int mlx5e_modify_sq(struct mlx5_core_dev *mdev, u32 sqn,
|
D | en_txrx.c | 81 nopwqe = mlx5e_post_nop(wq, sq->sqn, &sq->pc); in mlx5e_trigger_irq()
|
D | en_tx.c | 252 cseg->qpn_ds = cpu_to_be32((sq->sqn << 8) | ds_cnt); in mlx5e_txwqe_complete() 413 sq->cq.mcq.cqn, ci, sq->sqn, in mlx5e_dump_error_cqe()
|
D | eswitch_offloads.c | 549 u32 sqn) in mlx5_eswitch_add_send_to_vport_rule() argument 564 MLX5_SET(fte_match_set_misc, misc, source_sqn, sqn); in mlx5_eswitch_add_send_to_vport_rule()
|
D | en_rx.c | 480 mlx5e_post_nop(wq, sq->sqn, &sq->pc); in mlx5e_fill_icosq_frag_edge()
|
D | en_rep.c | 482 sqs[num_sqs++] = c->sq[tc].sqn; in mlx5e_add_sqs_fwd_rules()
|
/Linux-v5.4/include/linux/mlx5/ |
D | transobj.h | 46 u32 *sqn); 47 int mlx5_core_modify_sq(struct mlx5_core_dev *dev, u32 sqn, u32 *in, int inlen); 48 void mlx5_core_destroy_sq(struct mlx5_core_dev *dev, u32 sqn); 49 int mlx5_core_query_sq(struct mlx5_core_dev *dev, u32 sqn, u32 *out); 50 int mlx5_core_query_sq_state(struct mlx5_core_dev *dev, u32 sqn, u8 *state); 84 u32 *sqn; member
|
D | eswitch.h | 65 u16 vport_num, u32 sqn);
|
D | mlx5_ifc.h | 4451 u8 sqn[0x18]; member 5949 u8 sqn[0x18]; member 6671 u8 sqn[0x18]; member 7336 u8 sqn[0x18]; member
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en/ |
D | reporter_tx.c | 19 sq->sqn, sq->cc, sq->pc); in mlx5e_wait_for_sq_flush() 28 sq->sqn, sq->cc, sq->pc); in mlx5e_reset_txqsq_cc_pc() 49 err = mlx5_core_query_sq_state(mdev, sq->sqn, &state); in mlx5e_tx_reporter_err_cqe_recover() 52 sq->sqn, err); in mlx5e_tx_reporter_err_cqe_recover() 70 err = mlx5e_health_sq_to_ready(sq->channel, sq->sqn); in mlx5e_tx_reporter_err_cqe_recover() 93 sprintf(err_str, "ERR CQE on SQ: 0x%x", sq->sqn); in mlx5e_reporter_tx_err_cqe() 123 sq->channel->ix, sq->sqn, sq->cq.mcq.cqn, sq->cc, sq->pc, in mlx5e_reporter_tx_timeout() 156 err = mlx5_core_query_sq_state(priv->mdev, sq->sqn, &state); in mlx5e_tx_reporter_build_diagnose_output() 176 err = devlink_fmsg_u32_pair_put(fmsg, "sqn", sq->sqn); in mlx5e_tx_reporter_build_diagnose_output()
|
D | health.c | 130 int mlx5e_health_sq_to_ready(struct mlx5e_channel *channel, u32 sqn) in mlx5e_health_sq_to_ready() argument 140 err = mlx5e_modify_sq(mdev, sqn, &msp); in mlx5e_health_sq_to_ready() 142 netdev_err(dev, "Failed to move sq 0x%x to reset\n", sqn); in mlx5e_health_sq_to_ready() 150 err = mlx5e_modify_sq(mdev, sqn, &msp); in mlx5e_health_sq_to_ready() 152 netdev_err(dev, "Failed to move sq 0x%x to ready\n", sqn); in mlx5e_health_sq_to_ready()
|
D | reporter_rx.c | 43 icosq->sqn, icosq->cc, icosq->pc); in mlx5e_wait_for_icosq_flush() 51 icosq->sqn, icosq->cc, icosq->pc); in mlx5e_reset_icosq_cc_pc() 69 err = mlx5_core_query_sq_state(mdev, icosq->sqn, &state); in mlx5e_rx_reporter_err_icosq_cqe_recover() 72 icosq->sqn, err); in mlx5e_rx_reporter_err_icosq_cqe_recover() 88 err = mlx5e_health_sq_to_ready(icosq->channel, icosq->sqn); in mlx5e_rx_reporter_err_icosq_cqe_recover() 113 sprintf(err_str, "ERR CQE on ICOSQ: 0x%x", icosq->sqn); in mlx5e_reporter_icosq_cqe_err() 214 icosq->channel->ix, icosq->sqn, rq->rqn, rq->cq.mcq.cqn); in mlx5e_reporter_rx_timeout() 252 err = mlx5_core_query_sq_state(priv->mdev, icosq->sqn, &icosq_hw_state); in mlx5e_rx_reporter_build_diagnose_output()
|
D | txrx.h | 50 mlx5e_post_nop(struct mlx5_wq_cyc *wq, u32 sqn, u16 *pc) in mlx5e_post_nop() argument 59 cseg->qpn_ds = cpu_to_be32((sqn << 8) | 0x01); in mlx5e_post_nop() 67 mlx5e_post_nop_fence(struct mlx5_wq_cyc *wq, u32 sqn, u16 *pc) in mlx5e_post_nop_fence() argument 76 cseg->qpn_ds = cpu_to_be32((sqn << 8) | 0x01); in mlx5e_post_nop_fence() 96 mlx5e_post_nop(wq, sq->sqn, &sq->pc); in mlx5e_fill_sq_frag_edge()
|
D | health.h | 42 int mlx5e_health_sq_to_ready(struct mlx5e_channel *channel, u32 sqn);
|
D | xdp.h | 155 mlx5e_post_nop(wq, sq->sqn, &sq->pc); in mlx5e_fill_xdpsq_frag_edge()
|
D | xdp.c | 216 cseg->qpn_ds = cpu_to_be32((sq->sqn << 8) | ds_count); in mlx5e_xdp_mpwqe_complete()
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
D | ktls_tx.c | 52 build_static_params(struct mlx5e_umr_wqe *wqe, u16 pc, u32 sqn, in build_static_params() argument 64 cseg->qpn_ds = cpu_to_be32((sqn << MLX5_WQE_CTRL_QPN_SHIFT) | in build_static_params() 86 build_progress_params(struct mlx5e_tx_wqe *wqe, u16 pc, u32 sqn, in build_progress_params() argument 98 cseg->qpn_ds = cpu_to_be32((sqn << MLX5_WQE_CTRL_QPN_SHIFT) | in build_progress_params() 141 build_static_params(umr_wqe, sq->pc, sq->sqn, priv_tx, fence); in post_static_params() 155 build_progress_params(wqe, sq->pc, sq->sqn, priv_tx, fence); in post_progress_params() 276 cseg->qpn_ds = cpu_to_be32((sq->sqn << 8) | ds_cnt); in tx_post_resync_dump() 323 mlx5e_post_nop_fence(wq, sq->sqn, &sq->pc); in tx_post_fence_nop()
|
/Linux-v5.4/drivers/net/ethernet/mellanox/mlx5/core/en/xsk/ |
D | tx.c | 59 nopwqe = mlx5e_post_nop(&sq->wq, sq->sqn, &sq->pc); in mlx5e_xsk_tx_post_err()
|
/Linux-v5.4/include/uapi/rdma/ |
D | mlx5-abi.h | 373 __u32 sqn; member
|
/Linux-v5.4/drivers/scsi/bnx2i/ |
D | bnx2i_hwi.c | 156 if (cq_db->sqn[0] && cq_db->sqn[0] != 0xFFFF) in bnx2i_arm_cq_event_coalescing() 176 cq_db->sqn[0] = cq_index; in bnx2i_arm_cq_event_coalescing() 1126 memset(cq_db->sqn, 0xFF, sizeof(cq_db->sqn[0]) * BNX2X_MAX_CQS); in bnx2i_alloc_qp_resc()
|
D | bnx2i.h | 542 u16 sqn[BNX2X_MAX_CQS]; member
|
/Linux-v5.4/drivers/infiniband/hw/mlx5/ |
D | devx.c | 374 MLX5_GET(query_sq_in, in, sqn)); in devx_get_obj_id() 378 MLX5_GET(modify_sq_in, in, sqn)); in devx_get_obj_id()
|