Home
last modified time | relevance | path

Searched refs:mcq (Results 1 – 25 of 26) sorted by relevance

12

/Linux-v4.19/drivers/net/ethernet/mellanox/mlx4/
Den_cq.c101 cq->mcq.set_ci_db = cq->wqres.db.db; in mlx4_en_activate_cq()
102 cq->mcq.arm_db = cq->wqres.db.db + 1; in mlx4_en_activate_cq()
103 *cq->mcq.set_ci_db = 0; in mlx4_en_activate_cq()
104 *cq->mcq.arm_db = 0; in mlx4_en_activate_cq()
143 cq->mcq.usage = MLX4_RES_USAGE_DRIVER; in mlx4_en_activate_cq()
145 &mdev->priv_uar, cq->wqres.db.dma, &cq->mcq, in mlx4_en_activate_cq()
150 cq->mcq.event = mlx4_en_cq_event; in mlx4_en_activate_cq()
154 cq->mcq.comp = mlx4_en_tx_irq; in mlx4_en_activate_cq()
160 cq->mcq.comp = mlx4_en_rx_irq; in mlx4_en_activate_cq()
202 mlx4_cq_free(priv->mdev->dev, &cq->mcq); in mlx4_en_deactivate_cq()
[all …]
Dcq.c63 struct mlx4_cq *mcq, *temp; in mlx4_cq_tasklet_cb() local
69 list_for_each_entry_safe(mcq, temp, &ctx->process_list, tasklet_ctx.list) { in mlx4_cq_tasklet_cb()
70 list_del_init(&mcq->tasklet_ctx.list); in mlx4_cq_tasklet_cb()
71 mcq->tasklet_ctx.comp(mcq); in mlx4_cq_tasklet_cb()
72 if (refcount_dec_and_test(&mcq->refcount)) in mlx4_cq_tasklet_cb()
73 complete(&mcq->free); in mlx4_cq_tasklet_cb()
Den_rx.c344 ring->cqn = priv->rx_cq[ring_ind]->mcq.cqn; in mlx4_en_activate_rx_rings()
681 index = cq->mcq.cons_index & ring->size_mask; in mlx4_en_process_rx_cq()
686 cq->mcq.cons_index & cq->size)) { in mlx4_en_process_rx_cq()
882 ++cq->mcq.cons_index; in mlx4_en_process_rx_cq()
883 index = (cq->mcq.cons_index) & ring->size_mask; in mlx4_en_process_rx_cq()
897 mlx4_cq_set_ci(&cq->mcq); in mlx4_en_process_rx_cq()
899 ring->cons = cq->mcq.cons_index; in mlx4_en_process_rx_cq()
909 void mlx4_en_rx_irq(struct mlx4_cq *mcq) in mlx4_en_rx_irq() argument
911 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); in mlx4_en_rx_irq()
Den_tx.c392 struct mlx4_cq *mcq = &cq->mcq; in mlx4_en_process_tx_cq() local
398 u32 cons_index = mcq->cons_index; in mlx4_en_process_tx_cq()
477 mcq->cons_index = cons_index; in mlx4_en_process_tx_cq()
478 mlx4_cq_set_ci(mcq); in mlx4_en_process_tx_cq()
501 void mlx4_en_tx_irq(struct mlx4_cq *mcq) in mlx4_en_tx_irq() argument
503 struct mlx4_en_cq *cq = container_of(mcq, struct mlx4_en_cq, mcq); in mlx4_en_tx_irq()
Dmlx4_en.h367 struct mlx4_cq mcq; member
700 void mlx4_en_tx_irq(struct mlx4_cq *mcq);
764 void mlx4_en_rx_irq(struct mlx4_cq *mcq);
Den_netdev.c1674 priv->rx_ring[i]->cqn = cq->mcq.cqn; in mlx4_en_start_port()
1726 cq->mcq.cqn, in mlx4_en_start_port()
/Linux-v4.19/drivers/infiniband/hw/mlx4/
Dcq.c89 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw()
94 struct mlx4_ib_cq *mcq = to_mcq(cq); in mlx4_ib_modify_cq() local
97 return mlx4_cq_modify(dev->dev, &mcq->mcq, cq_count, cq_period); in mlx4_ib_modify_cq()
225 cq->mcq.usage = MLX4_RES_USAGE_USER_VERBS; in mlx4_ib_create_cq()
231 cq->mcq.set_ci_db = cq->db.db; in mlx4_ib_create_cq()
232 cq->mcq.arm_db = cq->db.db + 1; in mlx4_ib_create_cq()
233 *cq->mcq.set_ci_db = 0; in mlx4_ib_create_cq()
234 *cq->mcq.arm_db = 0; in mlx4_ib_create_cq()
241 cq->mcq.usage = MLX4_RES_USAGE_DRIVER; in mlx4_ib_create_cq()
248 cq->db.dma, &cq->mcq, vector, 0, in mlx4_ib_create_cq()
[all …]
Dmlx4_ib.h122 struct mlx4_cq mcq; member
669 static inline struct mlx4_ib_cq *to_mibcq(struct mlx4_cq *mcq) in to_mibcq() argument
671 return container_of(mcq, struct mlx4_ib_cq, mcq); in to_mibcq()
Dqp.c867 struct mlx4_ib_cq *mcq; in create_qp_common() local
1169 mcq = to_mcq(init_attr->send_cq); in create_qp_common()
1170 list_add_tail(&qp->cq_send_list, &mcq->send_qp_list); in create_qp_common()
1171 mcq = to_mcq(init_attr->recv_cq); in create_qp_common()
1172 list_add_tail(&qp->cq_recv_list, &mcq->recv_qp_list); in create_qp_common()
1241 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx4_ib_lock_cqs()
1256 } else if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx4_ib_unlock_cqs()
2295 context->cqn_send = cpu_to_be32(send_cq->mcq.cqn); in __mlx4_ib_modify_qp()
2296 context->cqn_recv = cpu_to_be32(recv_cq->mcq.cqn); in __mlx4_ib_modify_qp()
4357 struct mlx4_ib_cq *mcq = to_mcq(cq); in handle_drain_completion() local
[all …]
Dmain.c3157 struct mlx4_cq *mcq; in mlx4_ib_handle_catas_error() local
3171 if (send_mcq->mcq.comp && in mlx4_ib_handle_catas_error()
3173 if (!send_mcq->mcq.reset_notify_added) { in mlx4_ib_handle_catas_error()
3174 send_mcq->mcq.reset_notify_added = 1; in mlx4_ib_handle_catas_error()
3175 list_add_tail(&send_mcq->mcq.reset_notify, in mlx4_ib_handle_catas_error()
3189 if (recv_mcq->mcq.comp && in mlx4_ib_handle_catas_error()
3191 if (!recv_mcq->mcq.reset_notify_added) { in mlx4_ib_handle_catas_error()
3192 recv_mcq->mcq.reset_notify_added = 1; in mlx4_ib_handle_catas_error()
3193 list_add_tail(&recv_mcq->mcq.reset_notify, in mlx4_ib_handle_catas_error()
3204 list_for_each_entry(mcq, &cq_notify_list, reset_notify) { in mlx4_ib_handle_catas_error()
[all …]
Dsrq.c182 to_mcq(init_attr->ext.cq)->mcq.cqn : 0; in mlx4_ib_create_srq()
/Linux-v4.19/drivers/infiniband/hw/mlx5/
Dcq.c46 static void mlx5_ib_cq_event(struct mlx5_core_cq *mcq, enum mlx5_event type) in mlx5_ib_cq_event() argument
48 struct mlx5_ib_cq *cq = container_of(mcq, struct mlx5_ib_cq, mcq); in mlx5_ib_cq_event()
55 type, mcq->cqn); in mlx5_ib_cq_event()
82 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in get_sw_cqe()
94 return get_sw_cqe(cq, cq->mcq.cons_index); in next_cqe_sw()
531 cqe64 = (cq->mcq.cqe_sz == 64) ? cqe : cqe + 64; in mlx5_poll_one()
533 ++cq->mcq.cons_index; in mlx5_poll_one()
590 "Requestor" : "Responder", cq->mcq.cqn); in mlx5_poll_one()
626 cq->mcq.cqn, mr->sig->err_item.key, in mlx5_poll_one()
651 cq->mcq.cqn); in poll_soft_wc()
[all …]
Dqp.c1853 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(devr->c0)->mcq.cqn); in create_qp_common()
1854 MLX5_SET(qpc, qpc, cqn_snd, to_mcq(devr->c0)->mcq.cqn); in create_qp_common()
1859 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(devr->c0)->mcq.cqn); in create_qp_common()
1874 MLX5_SET(qpc, qpc, cqn_snd, to_mcq(init_attr->send_cq)->mcq.cqn); in create_qp_common()
1877 MLX5_SET(qpc, qpc, cqn_rcv, to_mcq(init_attr->recv_cq)->mcq.cqn); in create_qp_common()
1964 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs()
1968 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_lock_cqs()
1994 if (send_cq->mcq.cqn < recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs()
1997 } else if (send_cq->mcq.cqn == recv_cq->mcq.cqn) { in mlx5_ib_unlock_cqs()
2199 MLX5_SET(dctc, dctc, cqn, to_mcq(attr->recv_cq)->mcq.cqn); in mlx5_ib_create_dct()
[all …]
Dmlx5_ib.h488 struct mlx5_core_cq mcq; member
911 static inline struct mlx5_ib_cq *to_mibcq(struct mlx5_core_cq *mcq) in to_mibcq() argument
913 return container_of(mcq, struct mlx5_ib_cq, mcq); in to_mibcq()
Dsrq.c323 in.cqn = to_mcq(init_attr->ext.cq)->mcq.cqn; in mlx5_ib_create_srq()
325 in.cqn = to_mcq(dev->devr.c0)->mcq.cqn; in mlx5_ib_create_srq()
Dmain.c4124 struct mlx5_core_cq *mcq; in mlx5_ib_handle_internal_error() local
4139 if (send_mcq->mcq.comp && in mlx5_ib_handle_internal_error()
4141 if (!send_mcq->mcq.reset_notify_added) { in mlx5_ib_handle_internal_error()
4142 send_mcq->mcq.reset_notify_added = 1; in mlx5_ib_handle_internal_error()
4143 list_add_tail(&send_mcq->mcq.reset_notify, in mlx5_ib_handle_internal_error()
4156 if (recv_mcq->mcq.comp && in mlx5_ib_handle_internal_error()
4158 if (!recv_mcq->mcq.reset_notify_added) { in mlx5_ib_handle_internal_error()
4159 recv_mcq->mcq.reset_notify_added = 1; in mlx5_ib_handle_internal_error()
4160 list_add_tail(&recv_mcq->mcq.reset_notify, in mlx5_ib_handle_internal_error()
4173 list_for_each_entry(mcq, &cq_armed_list, reset_notify) { in mlx5_ib_handle_internal_error()
[all …]
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/
Den_txrx.c127 void mlx5e_completion_event(struct mlx5_core_cq *mcq) in mlx5e_completion_event() argument
129 struct mlx5e_cq *cq = container_of(mcq, struct mlx5e_cq, mcq); in mlx5e_completion_event()
136 void mlx5e_cq_error_event(struct mlx5_core_cq *mcq, enum mlx5_event event) in mlx5e_cq_error_event() argument
138 struct mlx5e_cq *cq = container_of(mcq, struct mlx5e_cq, mcq); in mlx5e_cq_error_event()
143 __func__, mcq->cqn, event); in mlx5e_cq_error_event()
Den_dim.c38 struct mlx5_core_dev *mdev, struct mlx5_core_cq *mcq) in mlx5e_complete_dim_work() argument
40 mlx5_core_modify_cq_moderation(mdev, mcq, moder.usec, moder.pkts); in mlx5e_complete_dim_work()
51 mlx5e_complete_dim_work(dim, cur_moder, rq->mdev, &rq->cq.mcq); in mlx5e_rx_dim_work()
61 mlx5e_complete_dim_work(dim, cur_moder, sq->cq.mdev, &sq->cq.mcq); in mlx5e_tx_dim_work()
Dcq.c50 struct mlx5_core_cq *mcq; in mlx5_cq_tasklet_cb() local
57 list_for_each_entry_safe(mcq, temp, &ctx->process_list, in mlx5_cq_tasklet_cb()
59 list_del_init(&mcq->tasklet_ctx.list); in mlx5_cq_tasklet_cb()
60 mcq->tasklet_ctx.comp(mcq); in mlx5_cq_tasklet_cb()
61 mlx5_cq_put(mcq); in mlx5_cq_tasklet_cb()
Den.h310 struct mlx5_core_cq mcq; member
735 void mlx5e_completion_event(struct mlx5_core_cq *mcq);
736 void mlx5e_cq_error_event(struct mlx5_core_cq *mcq, enum mlx5_event event);
885 struct mlx5_core_cq *mcq; in mlx5e_cq_arm() local
887 mcq = &cq->mcq; in mlx5e_cq_arm()
888 mlx5_cq_arm(mcq, MLX5_CQ_DB_REQ_NOT, mcq->uar->map, cq->wq.cc); in mlx5e_cq_arm()
Den_main.c748 MLX5_SET(rqc, rqc, cqn, rq->cq.mcq.cqn); in mlx5e_create_rq()
1308 csp.cqn = sq->cq.mcq.cqn; in mlx5e_open_txqsq()
1507 csp.cqn = sq->cq.mcq.cqn; in mlx5e_open_icosq()
1553 csp.cqn = sq->cq.mcq.cqn; in mlx5e_open_xdpsq()
1607 struct mlx5_core_cq *mcq = &cq->mcq; in mlx5e_alloc_cq_common() local
1620 mcq->cqe_sz = 64; in mlx5e_alloc_cq_common()
1621 mcq->set_ci_db = cq->wq_ctrl.db.db; in mlx5e_alloc_cq_common()
1622 mcq->arm_db = cq->wq_ctrl.db.db + 1; in mlx5e_alloc_cq_common()
1623 *mcq->set_ci_db = 0; in mlx5e_alloc_cq_common()
1624 *mcq->arm_db = 0; in mlx5e_alloc_cq_common()
[all …]
Den_ethtool.c433 &c->sq[tc].cq.mcq, in mlx5e_set_priv_channels_coalesce()
438 mlx5_core_modify_cq_moderation(mdev, &c->rq.cq.mcq, in mlx5e_set_priv_channels_coalesce()
Den_tx.c463 sq->cq.mcq.cqn, ci, sq->sqn, err_cqe->syndrome, in mlx5e_dump_error_cqe()
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/fpga/
Dconn.c361 mlx5_cq_arm(&conn->cq.mcq, MLX5_CQ_DB_REQ_NOT, in mlx5_fpga_conn_arm_cq()
365 static void mlx5_fpga_conn_cq_event(struct mlx5_core_cq *mcq, in mlx5_fpga_conn_cq_event() argument
370 conn = container_of(mcq, struct mlx5_fpga_conn, cq.mcq); in mlx5_fpga_conn_cq_event()
371 mlx5_fpga_warn(conn->fdev, "CQ event %u on CQ #%u\n", event, mcq->cqn); in mlx5_fpga_conn_cq_event()
417 static void mlx5_fpga_conn_cq_complete(struct mlx5_core_cq *mcq) in mlx5_fpga_conn_cq_complete() argument
421 conn = container_of(mcq, struct mlx5_fpga_conn, cq.mcq); in mlx5_fpga_conn_cq_complete()
479 err = mlx5_core_create_cq(mdev, &conn->cq.mcq, in, inlen); in mlx5_fpga_conn_create_cq()
485 conn->cq.mcq.cqe_sz = 64; in mlx5_fpga_conn_create_cq()
486 conn->cq.mcq.set_ci_db = conn->cq.wq_ctrl.db.db; in mlx5_fpga_conn_create_cq()
487 conn->cq.mcq.arm_db = conn->cq.wq_ctrl.db.db + 1; in mlx5_fpga_conn_create_cq()
[all …]
Dconn.h58 struct mlx5_core_cq mcq; member

12