Lines Matching refs:MLX5_CAP_GEN

251 	if (cap->max_recv_wr > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz)))  in set_rq_size()
277 if (wqe_size > MLX5_CAP_GEN(dev->mdev, max_wqe_sz_rq)) { in set_rq_size()
280 MLX5_CAP_GEN(dev->mdev, in set_rq_size()
404 if (wqe_size > MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)) { in calc_sq_size()
406 wqe_size, MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)); in calc_sq_size()
419 if (qp->sq.wqe_cnt > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz))) { in calc_sq_size()
423 1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz)); in calc_sq_size()
446 if (desc_sz > MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)) { in set_user_buf_size()
448 desc_sz, MLX5_CAP_GEN(dev->mdev, max_wqe_sz_sq)); in set_user_buf_size()
460 if (qp->sq.wqe_cnt > (1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz))) { in set_user_buf_size()
463 1 << MLX5_CAP_GEN(dev->mdev, log_max_qp_sz)); in set_user_buf_size()
943 qp->bf.buf_size = (1 << MLX5_CAP_GEN(dev->mdev, log_bf_reg_size)) / 2; in create_kernel_qp()
1126 if (MLX5_CAP_GEN(dev->mdev, eth_net_offloads) && in create_raw_packet_qp_sq()
1639 if (!MLX5_CAP_GEN(mdev, block_lb_mc)) { in create_qp_common()
1651 if (!MLX5_CAP_GEN(mdev, cd)) { in create_qp_common()
1665 if (!MLX5_CAP_GEN(mdev, ipoib_basic_offloads)) { in create_qp_common()
1675 if (!MLX5_CAP_GEN(dev->mdev, eth_net_offloads) || in create_qp_common()
1687 if (!(MLX5_CAP_GEN(dev->mdev, eth_net_offloads) && in create_qp_common()
1718 (MLX5_CAP_GEN(dev->mdev, port_type) != in create_qp_common()
1748 1 << MLX5_CAP_GEN(mdev, log_max_qp_sz); in create_qp_common()
1882 if (MLX5_CAP_GEN(mdev, cqe_version) == MLX5_CQE_VERSION_V1) in create_qp_common()
1893 if (!MLX5_CAP_GEN(dev->mdev, end_pad)) { in create_qp_common()
2241 if (!MLX5_CAP_GEN(dev->mdev, dct)) { in set_mlx_qp_type()
2306 if (!MLX5_CAP_GEN(dev->mdev, xrc)) { in mlx5_ib_create_qp()
2454 MLX5_CAP_GEN(dev->mdev, stat_rate_support))) in ib_rate_to_mlx5()
2751 if (MLX5_CAP_GEN(dev->mdev, modify_rq_counter_set_id)) { in modify_raw_packet_qp_rq()
3027 (u8)MLX5_CAP_GEN(dev->mdev, log_max_msg); in __mlx5_ib_modify_qp()
3300 attr->port_num > MLX5_CAP_GEN(dev->mdev, num_ports)) { in mlx5_ib_modify_dct()
3458 (1 << MLX5_CAP_GEN(dev->mdev, log_max_ra_res_qp))) { in mlx5_ib_modify_qp()
3466 (1 << MLX5_CAP_GEN(dev->mdev, log_max_ra_req_qp))) { in mlx5_ib_modify_qp()
3703 MLX5_CAP_GEN(dev->mdev, umr_modify_entity_size_disabled)) || in umr_check_mkey_mask()
3705 MLX5_CAP_GEN(dev->mdev, umr_modify_atomic_disabled))) in umr_check_mkey_mask()
4598 err = set_reg_umr_segment(dev, seg, wr, !!(MLX5_CAP_GEN(mdev, atomic))); in _mlx5_ib_post_send()
5179 if (!MLX5_CAP_GEN(dev->mdev, xrc)) in mlx5_ib_alloc_xrcd()
5283 if (!MLX5_CAP_GEN(dev->mdev, end_pad)) { in create_rq()
5306 has_net_offloads = MLX5_CAP_GEN(dev->mdev, eth_net_offloads); in create_rq()
5357 if (wq_init_attr->max_wr > (1 << MLX5_CAP_GEN(dev->mdev, log_max_wq_sz))) in set_user_rq_size()
5406 if (!MLX5_CAP_GEN(dev->mdev, striding_rq)) { in prepare_user_rq()
5546 MLX5_CAP_GEN(dev->mdev, log_max_rqt_size)) { in mlx5_ib_create_rwq_ind_table()
5549 MLX5_CAP_GEN(dev->mdev, log_max_rqt_size)); in mlx5_ib_create_rwq_ind_table()
5660 if (!(MLX5_CAP_GEN(dev->mdev, eth_net_offloads) && in mlx5_ib_modify_wq()
5681 if (MLX5_CAP_GEN(dev->mdev, modify_rq_counter_set_id)) { in mlx5_ib_modify_wq()