/Linux-v4.19/drivers/infiniband/hw/mthca/ |
D | mthca_srq.c | 75 return srq->queue.direct.buf + (n << srq->wqe_shift); in get_wqe() 77 return srq->queue.page_list[(n << srq->wqe_shift) >> PAGE_SHIFT].buf + in get_wqe() 78 ((n << srq->wqe_shift) & (PAGE_SIZE - 1)); in get_wqe() 102 context->wqe_base_ds = cpu_to_be64(1 << (srq->wqe_shift - 4)); in mthca_tavor_init_srq_context() 131 context->logstride_usrpage = cpu_to_be32((srq->wqe_shift - 4) << 29); in mthca_arbel_init_srq_context() 142 mthca_buf_free(dev, srq->max << srq->wqe_shift, &srq->queue, in mthca_free_srq_buf() 162 err = mthca_buf_alloc(dev, srq->max << srq->wqe_shift, in mthca_alloc_srq_buf() 182 next->nda_op = htonl(((i + 1) << srq->wqe_shift) | 1); in mthca_alloc_srq_buf() 189 (void *) scatter < wqe + (1 << srq->wqe_shift); in mthca_alloc_srq_buf() 227 srq->wqe_shift = ilog2(ds); in mthca_alloc_srq() [all …]
|
D | mthca_qp.c | 210 return qp->queue.direct.buf + (n << qp->rq.wqe_shift); in get_recv_wqe() 212 return qp->queue.page_list[(n << qp->rq.wqe_shift) >> PAGE_SHIFT].buf + in get_recv_wqe() 213 ((n << qp->rq.wqe_shift) & (PAGE_SIZE - 1)); in get_recv_wqe() 220 (n << qp->sq.wqe_shift); in get_send_wqe() 223 (n << qp->sq.wqe_shift)) >> in get_send_wqe() 225 ((qp->send_wqe_offset + (n << qp->sq.wqe_shift)) & in get_send_wqe() 612 qp_context->rq_size_stride |= qp->rq.wqe_shift - 4; in __mthca_modify_qp() 616 qp_context->sq_size_stride |= qp->sq.wqe_shift - 4; in __mthca_modify_qp() 963 1 << qp->sq.wqe_shift)); in mthca_adjust_qp_caps() 970 (min(dev->limits.max_desc_sz, 1 << qp->rq.wqe_shift) - in mthca_adjust_qp_caps() [all …]
|
D | mthca_provider.h | 229 int wqe_shift; member 255 int wqe_shift; member
|
D | mthca_cq.c | 541 >> wq->wqe_shift); in mthca_poll_one() 548 wqe_index = wqe >> srq->wqe_shift; in mthca_poll_one() 555 wqe_index = wqe >> wq->wqe_shift; in mthca_poll_one()
|
/Linux-v4.19/drivers/infiniband/hw/hns/ |
D | hns_roce_qp.c | 325 hr_qp->rq.wqe_shift = in hns_roce_set_rq_size() 328 hr_qp->rq.wqe_shift = in hns_roce_set_rq_size() 364 hr_qp->sq.wqe_shift = ucmd->log_sq_stride; in hns_roce_set_user_sq_size() 380 hr_qp->rq.wqe_shift), PAGE_SIZE) + in hns_roce_set_user_sq_size() 382 hr_qp->sq.wqe_shift), PAGE_SIZE); in hns_roce_set_user_sq_size() 386 hr_qp->sq.wqe_shift), PAGE_SIZE); in hns_roce_set_user_sq_size() 390 hr_qp->rq.wqe_shift), page_size) + in hns_roce_set_user_sq_size() 394 hr_qp->sq.wqe_shift), page_size); in hns_roce_set_user_sq_size() 400 hr_qp->sq.wqe_shift), in hns_roce_set_user_sq_size() 409 hr_qp->sq.wqe_shift), in hns_roce_set_user_sq_size() [all …]
|
D | hns_roce_device.h | 342 int wqe_shift;/* WQE size */ member
|
/Linux-v4.19/include/linux/mlx5/ |
D | srq.h | 48 u32 wqe_shift; member
|
D | driver.h | 475 int wqe_shift; member
|
/Linux-v4.19/drivers/infiniband/hw/mlx4/ |
D | srq.c | 43 return mlx4_buf_offset(&srq->buf, n << srq->msrq.wqe_shift); in get_wqe() 104 srq->msrq.wqe_shift = ilog2(desc_size); in mlx4_ib_create_srq() 286 mlx4_buf_free(dev->dev, msrq->msrq.max << msrq->msrq.wqe_shift, in mlx4_ib_destroy_srq()
|
D | qp.c | 197 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); in get_recv_wqe() 202 return get_wqe(qp, qp->sq.offset + (n << qp->sq.wqe_shift)); in get_send_wqe() 352 qp->rq.wqe_shift = ilog2(max_t(u32, wqe_size, inl_recv_sz)); in set_rq_size() 398 qp->sq.wqe_shift = ilog2(roundup_pow_of_two(s)); in set_kernel_sq_size() 404 qp->sq_spare_wqes = (2048 >> qp->sq.wqe_shift) + 1; in set_kernel_sq_size() 410 (1 << qp->sq.wqe_shift)) - in set_kernel_sq_size() 414 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_kernel_sq_size() 415 (qp->sq.wqe_cnt << qp->sq.wqe_shift); in set_kernel_sq_size() 416 if (qp->rq.wqe_shift > qp->sq.wqe_shift) { in set_kernel_sq_size() 418 qp->sq.offset = qp->rq.wqe_cnt << qp->rq.wqe_shift; in set_kernel_sq_size() [all …]
|
D | mlx4_ib.h | 179 int wqe_shift; member
|
/Linux-v4.19/drivers/infiniband/hw/mlx5/ |
D | srq.c | 47 return mlx5_buf_offset(&srq->buf, n << srq->msrq.wqe_shift); in get_wqe() 281 srq->msrq.wqe_shift = ilog2(desc_size); in mlx5_ib_create_srq() 301 in.wqe_shift = srq->msrq.wqe_shift - 4; in mlx5_ib_create_srq()
|
D | qp.c | 117 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); in get_recv_wqe() 162 offset = wq->offset + ((wqe_index % wq->wqe_cnt) << wq->wqe_shift); in mlx5_ib_read_user_wqe() 163 wq_end = wq->offset + (wq->wqe_cnt << wq->wqe_shift); in mlx5_ib_read_user_wqe() 183 wqe_length = 1 << wq->wqe_shift; in mlx5_ib_read_user_wqe() 257 qp->rq.wqe_shift = 0; in set_rq_size() 265 qp->rq.wqe_shift = ucmd->rq_wqe_shift; in set_rq_size() 266 if ((1 << qp->rq.wqe_shift) / sizeof(struct mlx5_wqe_data_seg) < qp->wq_sig) in set_rq_size() 268 qp->rq.max_gs = (1 << qp->rq.wqe_shift) / sizeof(struct mlx5_wqe_data_seg) - qp->wq_sig; in set_rq_size() 284 qp->rq.wqe_shift = ilog2(wqe_size); in set_rq_size() 285 qp->rq.max_gs = (1 << qp->rq.wqe_shift) / sizeof(struct mlx5_wqe_data_seg) - qp->wq_sig; in set_rq_size() [all …]
|
D | mlx5_ib.h | 270 int wqe_shift; member 306 u32 wqe_shift; member
|
D | odp.c | 973 int wqe_size = 1 << wq->wqe_shift; in mlx5_ib_mr_responder_pfault_handler()
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx5/core/ |
D | srq.c | 70 u32 log_rq_stride = in->wqe_shift; in get_pas_size() 86 MLX5_SET(wq, wq, log_wq_stride, in->wqe_shift + 4); in set_wq() 99 MLX5_SET(srqc, srqc, log_rq_stride, in->wqe_shift); in set_srqc() 114 in->wqe_shift = MLX5_GET(wq, wq, log_wq_stride) - 4; in get_wq() 127 in->wqe_shift = MLX5_GET(srqc, srqc, log_rq_stride); in get_srqc()
|
/Linux-v4.19/drivers/net/ethernet/mellanox/mlx4/ |
D | srq.c | 190 srq_context->logstride = srq->wqe_shift - 4; in mlx4_srq_alloc()
|
/Linux-v4.19/include/linux/mlx4/ |
D | device.h | 780 int wqe_shift; member
|