/Linux-v4.19/drivers/net/ethernet/mellanox/mlx4/ |
D | cq.c | 173 struct mlx4_cq_context *cq_context; in mlx4_cq_modify() local 180 cq_context = mailbox->buf; in mlx4_cq_modify() 181 cq_context->cq_max_count = cpu_to_be16(count); in mlx4_cq_modify() 182 cq_context->cq_period = cpu_to_be16(period); in mlx4_cq_modify() 195 struct mlx4_cq_context *cq_context; in mlx4_cq_resize() local 203 cq_context = mailbox->buf; in mlx4_cq_resize() 204 cq_context->logsize_usrpage = cpu_to_be32(ilog2(entries) << 24); in mlx4_cq_resize() 205 cq_context->log_page_size = mtt->page_shift - 12; in mlx4_cq_resize() 207 cq_context->mtt_base_addr_h = mtt_addr >> 32; in mlx4_cq_resize() 208 cq_context->mtt_base_addr_l = cpu_to_be32(mtt_addr & 0xffffffff); in mlx4_cq_resize() [all …]
|
/Linux-v4.19/drivers/infiniband/hw/mthca/ |
D | mthca_cq.c | 235 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in mthca_cq_completion() 261 cq->ibcq.event_handler(&event, cq->ibcq.cq_context); in mthca_cq_event() 778 struct mthca_cq_context *cq_context; in mthca_init_cq() local 814 cq_context = mailbox->buf; in mthca_init_cq() 827 memset(cq_context, 0, sizeof *cq_context); in mthca_init_cq() 828 cq_context->flags = cpu_to_be32(MTHCA_CQ_STATUS_OK | in mthca_init_cq() 831 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq() 833 cq_context->logsize_usrpage |= cpu_to_be32(ctx->uar.index); in mthca_init_cq() 835 cq_context->logsize_usrpage |= cpu_to_be32(dev->driver_uar.index); in mthca_init_cq() 836 cq_context->error_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn); in mthca_init_cq() [all …]
|
/Linux-v4.19/net/smc/ |
D | smc_wr.h | 93 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context); 101 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context);
|
D | smc_wr.c | 137 void smc_wr_tx_cq_handler(struct ib_cq *ib_cq, void *cq_context) in smc_wr_tx_cq_handler() argument 139 struct smc_ib_device *dev = (struct smc_ib_device *)cq_context; in smc_wr_tx_cq_handler() 405 void smc_wr_rx_cq_handler(struct ib_cq *ib_cq, void *cq_context) in smc_wr_rx_cq_handler() argument 407 struct smc_ib_device *dev = (struct smc_ib_device *)cq_context; in smc_wr_rx_cq_handler()
|
/Linux-v4.19/drivers/infiniband/sw/rxe/ |
D | rxe_cq.c | 81 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in rxe_send_complete() 145 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rxe_cq_post()
|
/Linux-v4.19/drivers/infiniband/core/ |
D | uverbs_std_types_cq.c | 41 struct ib_uverbs_event_queue *ev_queue = cq->cq_context; in uverbs_free_cq() 127 cq->cq_context = ev_file ? &ev_file->ev_queue : NULL; in UVERBS_HANDLER()
|
D | cq.c | 155 cq->cq_context = private; in __ib_alloc_cq()
|
D | uverbs.h | 241 void ib_uverbs_comp_handler(struct ib_cq *cq, void *cq_context);
|
/Linux-v4.19/drivers/infiniband/hw/cxgb3/ |
D | iwch_ev.c | 100 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 178 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in iwch_ev_dispatch()
|
D | iwch_qp.c | 748 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in __flush_qp() 762 (*schp->ibcq.comp_handler)(&schp->ibcq, schp->ibcq.cq_context); in __flush_qp() 784 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in flush_qp() 790 schp->ibcq.cq_context); in flush_qp()
|
/Linux-v4.19/drivers/infiniband/hw/hns/ |
D | hns_roce_hw_v2.c | 1905 struct hns_roce_v2_cq_context *cq_context; in hns_roce_v2_write_cqc() local 1907 cq_context = mb_buf; in hns_roce_v2_write_cqc() 1908 memset(cq_context, 0, sizeof(*cq_context)); in hns_roce_v2_write_cqc() 1910 roce_set_field(cq_context->byte_4_pg_ceqn, V2_CQC_BYTE_4_CQ_ST_M, in hns_roce_v2_write_cqc() 1912 roce_set_field(cq_context->byte_4_pg_ceqn, V2_CQC_BYTE_4_ARM_ST_M, in hns_roce_v2_write_cqc() 1914 roce_set_field(cq_context->byte_4_pg_ceqn, V2_CQC_BYTE_4_SHIFT_M, in hns_roce_v2_write_cqc() 1916 roce_set_field(cq_context->byte_4_pg_ceqn, V2_CQC_BYTE_4_CEQN_M, in hns_roce_v2_write_cqc() 1918 cq_context->byte_4_pg_ceqn = cpu_to_le32(cq_context->byte_4_pg_ceqn); in hns_roce_v2_write_cqc() 1920 roce_set_field(cq_context->byte_8_cqn, V2_CQC_BYTE_8_CQN_M, in hns_roce_v2_write_cqc() 1923 cq_context->cqe_cur_blk_addr = (u32)(mtts[0] >> PAGE_ADDR_SHIFT); in hns_roce_v2_write_cqc() [all …]
|
D | hns_roce_hw_v1.c | 742 free_mr->mr_free_cq->ib_cq.cq_context = NULL; in hns_roce_v1_rsv_lp_qp() 2111 struct hns_roce_cq_context *cq_context = NULL; in hns_roce_v1_write_cqc() local 2120 cq_context = mb_buf; in hns_roce_v1_write_cqc() 2121 memset(cq_context, 0, sizeof(*cq_context)); in hns_roce_v1_write_cqc() 2129 roce_set_field(cq_context->cqc_byte_4, in hns_roce_v1_write_cqc() 2132 roce_set_field(cq_context->cqc_byte_4, CQ_CONTEXT_CQC_BYTE_4_CQN_M, in hns_roce_v1_write_cqc() 2135 cq_context->cq_bt_l = cpu_to_le32((u32)dma_handle); in hns_roce_v1_write_cqc() 2137 roce_set_field(cq_context->cqc_byte_12, in hns_roce_v1_write_cqc() 2141 roce_set_field(cq_context->cqc_byte_12, in hns_roce_v1_write_cqc() 2145 roce_set_field(cq_context->cqc_byte_12, CQ_CONTEXT_CQC_BYTE_12_CEQN_M, in hns_roce_v1_write_cqc() [all …]
|
D | hns_roce_cq.c | 45 ibcq->comp_handler(ibcq, ibcq->cq_context); in hns_roce_ib_cq_comp() 71 ibcq->event_handler(&event, ibcq->cq_context); in hns_roce_ib_cq_event()
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/ |
D | cq.c | 91 int eqn = MLX5_GET(cqc, MLX5_ADDR_OF(create_cq_in, in, cq_context), c_eqn); in mlx5_core_create_cq() 212 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_core_modify_cq_moderation()
|
D | debugfs.c | 407 ctx = MLX5_ADDR_OF(query_cq_out, out, cq_context); in cq_read_field()
|
/Linux-v4.19/drivers/infiniband/hw/cxgb4/ |
D | ev.c | 115 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in post_qp_event() 235 (*chp->ibcq.comp_handler)(&chp->ibcq, chp->ibcq.cq_context); in c4iw_ev_handler()
|
D | qp.c | 1032 schp->ibcq.cq_context); in complete_sq_drain_wr() 1083 rchp->ibcq.cq_context); in complete_rq_drain_wr() 1660 rchp->ibcq.cq_context); in __flush_qp() 1667 rchp->ibcq.cq_context); in __flush_qp() 1673 schp->ibcq.cq_context); in __flush_qp() 1697 (*rchp->ibcq.comp_handler)(&rchp->ibcq, rchp->ibcq.cq_context); in flush_qp() 1703 schp->ibcq.cq_context); in flush_qp()
|
/Linux-v4.19/drivers/infiniband/hw/mlx5/ |
D | cq.c | 43 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx5_ib_cq_comp() 63 ibcq->event_handler(&event, ibcq->cq_context); in mlx5_ib_cq_event() 834 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_user() 947 cqc = MLX5_ADDR_OF(create_cq_in, *cqb, cq_context); in create_cq_kernel() 975 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in notify_soft_wc_handler() 1041 cqc = MLX5_ADDR_OF(create_cq_in, cqb, cq_context); in mlx5_ib_create_cq() 1402 cqc = MLX5_ADDR_OF(modify_cq_in, in, cq_context); in mlx5_ib_resize_cq()
|
D | gsi.c | 98 struct mlx5_ib_gsi_qp *gsi = cq->cq_context; in handle_single_completion()
|
/Linux-v4.19/drivers/infiniband/sw/rdmavt/ |
D | cq.c | 94 cq->ibcq.event_handler(&ev, cq->ibcq.cq_context); in rvt_cq_enter() 159 cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context); in send_complete()
|
/Linux-v4.19/net/9p/ |
D | trans_rdma.c | 308 struct p9_client *client = cq->cq_context; in recv_done() 358 struct p9_client *client = cq->cq_context; in send_done()
|
/Linux-v4.19/drivers/infiniband/hw/i40iw/ |
D | i40iw_hw.c | 177 i40iwcq->ibcq.cq_context); in i40iw_iwarp_ce_handler() 387 iwcq->ibcq.event_handler(&ibevent, iwcq->ibcq.cq_context); in i40iw_process_aeq()
|
/Linux-v4.19/drivers/infiniband/hw/qedr/ |
D | qedr_roce_cm.c | 93 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_tx_packet() 122 (*cq->ibcq.comp_handler) (&cq->ibcq, cq->ibcq.cq_context); in qedr_ll2_complete_rx_packet()
|
D | main.c | 498 (&cq->ibcq, cq->ibcq.cq_context); in qedr_irq_handler() 738 ibcq->event_handler(&event, ibcq->cq_context); in qedr_affiliated_event()
|
/Linux-v4.19/drivers/infiniband/hw/mlx4/ |
D | cq.c | 45 ibcq->comp_handler(ibcq, ibcq->cq_context); in mlx4_ib_cq_comp() 64 ibcq->event_handler(&event, ibcq->cq_context); in mlx4_ib_cq_event()
|