Lines Matching refs:access_flags

56 				     u64 iova, int access_flags,
304 set_mkc_access_pd_addr_fields(mkc, ent->rb_key.access_flags, 0, in set_cache_mkc()
672 res = key1.access_flags - key2.access_flags; in cache_ent_key_cmp()
739 smallest->rb_key.access_flags == rb_key.access_flags && in mkey_cache_ent_from_rb_key()
747 int access_flags) in _mlx5_mr_cache_alloc() argument
783 int access_flags) in get_unchangeable_access_flags() argument
787 if ((access_flags & IB_ACCESS_REMOTE_ATOMIC) && in get_unchangeable_access_flags()
792 if ((access_flags & IB_ACCESS_RELAXED_ORDERING) && in get_unchangeable_access_flags()
797 if ((access_flags & IB_ACCESS_RELAXED_ORDERING) && in get_unchangeable_access_flags()
807 int access_flags, int access_mode, in mlx5_mr_cache_alloc() argument
813 .access_flags = get_unchangeable_access_flags(dev, access_flags) in mlx5_mr_cache_alloc()
820 return _mlx5_mr_cache_alloc(dev, ent, access_flags); in mlx5_mr_cache_alloc()
1128 u64 length, int access_flags, u64 iova) in set_mr_fields() argument
1135 mr->access_flags = access_flags; in set_mr_fields()
1151 int access_flags) in alloc_cacheable_mr() argument
1170 rb_key.ats = mlx5_umem_needs_ats(dev, umem, access_flags); in alloc_cacheable_mr()
1171 rb_key.access_flags = get_unchangeable_access_flags(dev, access_flags); in alloc_cacheable_mr()
1179 mr = reg_create(pd, umem, iova, access_flags, page_size, false); in alloc_cacheable_mr()
1187 mr = _mlx5_mr_cache_alloc(dev, ent, access_flags); in alloc_cacheable_mr()
1194 set_mr_fields(dev, mr, umem->length, access_flags, iova); in alloc_cacheable_mr()
1204 u64 iova, int access_flags, in reg_create() argument
1223 mr->access_flags = access_flags; in reg_create()
1237 if (WARN_ON(access_flags & IB_ACCESS_ON_DEMAND)) { in reg_create()
1251 set_mkc_access_pd_addr_fields(mkc, access_flags, iova, in reg_create()
1262 if (mlx5_umem_needs_ats(dev, umem, access_flags)) in reg_create()
1277 set_mr_fields(dev, mr, umem->length, access_flags, iova); in reg_create()
1364 if (attr->access_flags & ~MLX5_IB_DM_MEMIC_ALLOWED_ACCESS) in mlx5_ib_reg_dm_mr()
1373 if (attr->access_flags & ~MLX5_IB_DM_SW_ICM_ALLOWED_ACCESS) in mlx5_ib_reg_dm_mr()
1383 attr->access_flags, mode); in mlx5_ib_reg_dm_mr()
1387 u64 iova, int access_flags) in create_real_mr() argument
1396 mr = alloc_cacheable_mr(pd, umem, iova, access_flags); in create_real_mr()
1402 mr = reg_create(pd, umem, iova, access_flags, page_size, true); in create_real_mr()
1430 u64 iova, int access_flags, in create_user_odp_mr() argument
1450 mr = mlx5_ib_alloc_implicit_mr(to_mpd(pd), access_flags); in create_user_odp_mr()
1460 odp = ib_umem_odp_get(&dev->ib_dev, start, length, access_flags, in create_user_odp_mr()
1465 mr = alloc_cacheable_mr(pd, &odp->umem, iova, access_flags); in create_user_odp_mr()
1488 u64 iova, int access_flags, in mlx5_ib_reg_user_mr() argument
1498 start, iova, length, access_flags); in mlx5_ib_reg_user_mr()
1500 if (access_flags & IB_ACCESS_ON_DEMAND) in mlx5_ib_reg_user_mr()
1501 return create_user_odp_mr(pd, start, length, iova, access_flags, in mlx5_ib_reg_user_mr()
1503 umem = ib_umem_get(&dev->ib_dev, start, length, access_flags); in mlx5_ib_reg_user_mr()
1506 return create_real_mr(pd, umem, iova, access_flags); in mlx5_ib_reg_user_mr()
1530 int fd, int access_flags, in mlx5_ib_reg_user_mr_dmabuf() argument
1544 offset, virt_addr, length, fd, access_flags); in mlx5_ib_reg_user_mr_dmabuf()
1551 access_flags, in mlx5_ib_reg_user_mr_dmabuf()
1560 access_flags); in mlx5_ib_reg_user_mr_dmabuf()
1623 int access_flags, int flags, struct ib_umem *new_umem, in umr_rereg_pas() argument
1645 mr->access_flags = access_flags; in umr_rereg_pas()
1690 new_access_flags = mr->access_flags; in mlx5_ib_rereg_user_mr()
1698 if (can_use_umr_rereg_access(dev, mr->access_flags, in mlx5_ib_rereg_user_mr()
1733 can_use_umr_rereg_access(dev, mr->access_flags, new_access_flags)) { in mlx5_ib_rereg_user_mr()