Home
last modified time | relevance | path

Searched refs:resv (Results 1 – 25 of 203) sorted by relevance

123456789

/Linux-v5.15/fs/ocfs2/
Dreservations.c40 struct ocfs2_alloc_reservation *resv) in ocfs2_resv_window_bits() argument
45 if (!(resv->r_flags & OCFS2_RESV_FLAG_DIR)) { in ocfs2_resv_window_bits()
54 static inline unsigned int ocfs2_resv_end(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_end() argument
56 if (resv->r_len) in ocfs2_resv_end()
57 return resv->r_start + resv->r_len - 1; in ocfs2_resv_end()
58 return resv->r_start; in ocfs2_resv_end()
61 static inline int ocfs2_resv_empty(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_empty() argument
63 return !!(resv->r_len == 0); in ocfs2_resv_empty()
77 struct ocfs2_alloc_reservation *resv; in ocfs2_dump_resv() local
85 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_dump_resv()
[all …]
Dreservations.h54 void ocfs2_resv_init_once(struct ocfs2_alloc_reservation *resv);
57 void ocfs2_resv_set_type(struct ocfs2_alloc_reservation *resv,
71 struct ocfs2_alloc_reservation *resv);
127 struct ocfs2_alloc_reservation *resv,
146 struct ocfs2_alloc_reservation *resv,
/Linux-v5.15/fs/xfs/libxfs/
Dxfs_ag_resv.c133 struct xfs_ag_resv *resv; in __xfs_ag_resv_free() local
139 resv = xfs_perag_resv(pag, type); in __xfs_ag_resv_free()
141 pag->pag_mount->m_ag_max_usable += resv->ar_asked; in __xfs_ag_resv_free()
148 oldresv = resv->ar_orig_reserved; in __xfs_ag_resv_free()
150 oldresv = resv->ar_reserved; in __xfs_ag_resv_free()
152 resv->ar_reserved = 0; in __xfs_ag_resv_free()
153 resv->ar_asked = 0; in __xfs_ag_resv_free()
154 resv->ar_orig_reserved = 0; in __xfs_ag_resv_free()
185 struct xfs_ag_resv *resv; in __xfs_ag_resv_init() local
237 resv = xfs_perag_resv(pag, type); in __xfs_ag_resv_init()
[all …]
/Linux-v5.15/drivers/gpu/drm/ttm/
Dttm_bo.c98 dma_resv_assert_held(bo->base.resv); in ttm_bo_move_to_lru_tail()
139 dma_resv_assert_held(pos->first->base.resv); in ttm_bo_bulk_move_lru_tail()
140 dma_resv_assert_held(pos->last->base.resv); in ttm_bo_bulk_move_lru_tail()
154 dma_resv_assert_held(pos->first->base.resv); in ttm_bo_bulk_move_lru_tail()
155 dma_resv_assert_held(pos->last->base.resv); in ttm_bo_bulk_move_lru_tail()
236 if (bo->base.resv == &bo->base._resv) in ttm_bo_individualize_resv()
241 r = dma_resv_copy_fences(&bo->base._resv, bo->base.resv); in ttm_bo_individualize_resv()
252 bo->base.resv = &bo->base._resv; in ttm_bo_individualize_resv()
261 struct dma_resv *resv = &bo->base._resv; in ttm_bo_flush_all_fences() local
267 fobj = dma_resv_shared_list(resv); in ttm_bo_flush_all_fences()
[all …]
Dttm_execbuf_util.c42 dma_resv_unlock(bo->base.resv); in ttm_eu_backoff_reservation_reverse()
58 dma_resv_unlock(bo->base.resv); in ttm_eu_backoff_reservation()
107 ret = dma_resv_reserve_shared(bo->base.resv, in ttm_eu_reserve_buffers()
124 ret = dma_resv_reserve_shared(bo->base.resv, in ttm_eu_reserve_buffers()
159 dma_resv_add_shared_fence(bo->base.resv, fence); in ttm_eu_fence_buffer_objects()
161 dma_resv_add_excl_fence(bo->base.resv, fence); in ttm_eu_fence_buffer_objects()
163 dma_resv_unlock(bo->base.resv); in ttm_eu_fence_buffer_objects()
/Linux-v5.15/drivers/dma-buf/
Ddma-buf.c82 if (dmabuf->resv == (struct dma_resv *)&dmabuf[1]) in dma_buf_release()
83 dma_resv_fini(dmabuf->resv); in dma_buf_release()
213 struct dma_resv *resv; in dma_buf_poll() local
220 if (!dmabuf || !dmabuf->resv) in dma_buf_poll()
223 resv = dmabuf->resv; in dma_buf_poll()
232 seq = read_seqcount_begin(&resv->seq); in dma_buf_poll()
235 fobj = rcu_dereference(resv->fence); in dma_buf_poll()
240 fence_excl = dma_resv_excl_fence(resv); in dma_buf_poll()
241 if (read_seqcount_retry(&resv->seq, seq)) { in dma_buf_poll()
352 dma_resv_lock(dmabuf->resv, NULL); in dma_buf_set_name()
[all …]
/Linux-v5.15/drivers/gpu/drm/i915/
Ddma_resv_utils.c10 void dma_resv_prune(struct dma_resv *resv) in dma_resv_prune() argument
12 if (dma_resv_trylock(resv)) { in dma_resv_prune()
13 if (dma_resv_test_signaled(resv, true)) in dma_resv_prune()
14 dma_resv_add_excl_fence(resv, NULL); in dma_resv_prune()
15 dma_resv_unlock(resv); in dma_resv_prune()
/Linux-v5.15/drivers/gpu/drm/vgem/
Dvgem_fence.c131 struct dma_resv *resv; in vgem_fence_attach_ioctl() local
153 resv = obj->resv; in vgem_fence_attach_ioctl()
154 if (!dma_resv_test_signaled(resv, arg->flags & VGEM_FENCE_WRITE)) { in vgem_fence_attach_ioctl()
161 dma_resv_lock(resv, NULL); in vgem_fence_attach_ioctl()
163 dma_resv_add_excl_fence(resv, fence); in vgem_fence_attach_ioctl()
164 else if ((ret = dma_resv_reserve_shared(resv, 1)) == 0) in vgem_fence_attach_ioctl()
165 dma_resv_add_shared_fence(resv, fence); in vgem_fence_attach_ioctl()
166 dma_resv_unlock(resv); in vgem_fence_attach_ioctl()
/Linux-v5.15/drivers/gpu/drm/amd/amdgpu/
Damdgpu_dma_buf.c334 struct dma_resv *resv = dma_buf->resv; in amdgpu_dma_buf_create_obj() local
341 dma_resv_lock(resv, NULL); in amdgpu_dma_buf_create_obj()
351 ttm_bo_type_sg, resv, &gobj); in amdgpu_dma_buf_create_obj()
359 dma_resv_unlock(resv); in amdgpu_dma_buf_create_obj()
363 dma_resv_unlock(resv); in amdgpu_dma_buf_create_obj()
379 struct ww_acquire_ctx *ticket = dma_resv_locking_ctx(obj->resv); in amdgpu_dma_buf_move_notify()
398 struct dma_resv *resv = vm->root.bo->tbo.base.resv; in amdgpu_dma_buf_move_notify() local
405 r = dma_resv_lock(resv, ticket); in amdgpu_dma_buf_move_notify()
414 if (!dma_resv_trylock(resv)) in amdgpu_dma_buf_move_notify()
426 dma_resv_unlock(resv); in amdgpu_dma_buf_move_notify()
Damdgpu_vm.c375 if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv) in amdgpu_vm_bo_base_init()
665 if (abo->tbo.base.resv == vm->root.bo->tbo.base.resv) in amdgpu_vm_del_from_lru_notify()
920 struct dma_resv *resv; in amdgpu_vm_pt_create() local
946 bp.resv = vm->root.bo->tbo.base.resv; in amdgpu_vm_pt_create()
958 if (!bp.resv) in amdgpu_vm_pt_create()
959 WARN_ON(dma_resv_lock(bo->tbo.base.resv, in amdgpu_vm_pt_create()
961 resv = bp.resv; in amdgpu_vm_pt_create()
967 bp.resv = bo->tbo.base.resv; in amdgpu_vm_pt_create()
972 if (!resv) in amdgpu_vm_pt_create()
973 dma_resv_unlock(bo->tbo.base.resv); in amdgpu_vm_pt_create()
[all …]
Damdgpu_object.c249 bp.resv = NULL; in amdgpu_bo_create_reserved()
533 .resv = bp->resv in amdgpu_bo_create()
591 bp->resv, bp->destroy); in amdgpu_bo_create()
607 r = amdgpu_fill_buffer(bo, 0, bo->tbo.base.resv, &fence); in amdgpu_bo_create()
616 if (!bp->resv) in amdgpu_bo_create()
629 if (!bp->resv) in amdgpu_bo_create()
630 dma_resv_unlock(bo->tbo.base.resv); in amdgpu_bo_create()
801 r = dma_resv_wait_timeout(bo->tbo.base.resv, false, false, in amdgpu_bo_kmap()
1158 dma_resv_assert_held(bo->tbo.base.resv); in amdgpu_bo_get_tiling_flags()
1338 && bo->base.resv != &bo->base._resv); in amdgpu_bo_release_notify()
[all …]
Damdgpu_vm_cpu.c48 struct dma_resv *resv, in amdgpu_vm_cpu_prepare() argument
51 if (!resv) in amdgpu_vm_cpu_prepare()
54 return amdgpu_bo_sync_wait_resv(p->adev, resv, sync_mode, p->vm, true); in amdgpu_vm_cpu_prepare()
/Linux-v5.15/include/drm/ttm/
Dttm_bo_driver.h143 success = dma_resv_trylock(bo->base.resv); in ttm_bo_reserve()
148 ret = dma_resv_lock_interruptible(bo->base.resv, ticket); in ttm_bo_reserve()
150 ret = dma_resv_lock(bo->base.resv, ticket); in ttm_bo_reserve()
171 int ret = dma_resv_lock_slow_interruptible(bo->base.resv, in ttm_bo_reserve_slowpath()
177 dma_resv_lock_slow(bo->base.resv, ticket); in ttm_bo_reserve_slowpath()
220 dma_resv_unlock(bo->base.resv); in ttm_bo_unreserve()
Dttm_bo_api.h214 struct dma_resv *resv; member
399 struct sg_table *sg, struct dma_resv *resv,
442 struct sg_table *sg, struct dma_resv *resv,
562 dma_resv_assert_held(bo->base.resv); in ttm_bo_pin()
575 dma_resv_assert_held(bo->base.resv); in ttm_bo_unpin()
/Linux-v5.15/include/uapi/linux/
Dio_uring.h272 __u32 resv[3]; member
336 __u32 resv; member
342 __u32 resv; member
350 __u32 resv; member
356 __u32 resv; member
370 __u8 resv; member
378 __u16 resv; member
390 __u8 resv; member
/Linux-v5.15/drivers/gpu/drm/radeon/
Dradeon_prime.c47 struct dma_resv *resv = attach->dmabuf->resv; in radeon_gem_prime_import_sg_table() local
52 dma_resv_lock(resv, NULL); in radeon_gem_prime_import_sg_table()
54 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); in radeon_gem_prime_import_sg_table()
55 dma_resv_unlock(resv); in radeon_gem_prime_import_sg_table()
Dradeon_benchmark.c38 struct dma_resv *resv) in radeon_benchmark_do_move() argument
51 resv); in radeon_benchmark_do_move()
56 resv); in radeon_benchmark_do_move()
125 dobj->tbo.base.resv); in radeon_benchmark_move()
136 dobj->tbo.base.resv); in radeon_benchmark_move()
Dradeon_sync.c91 struct dma_resv *resv, in radeon_sync_resv() argument
101 f = dma_resv_excl_fence(resv); in radeon_sync_resv()
108 flist = dma_resv_shared_list(resv); in radeon_sync_resv()
114 dma_resv_held(resv)); in radeon_sync_resv()
/Linux-v5.15/net/sunrpc/
Dsvc.c1193 struct kvec *resv = &rqstp->rq_res.head[0]; in svc_generic_dispatch() local
1220 !procp->pc_encode(rqstp, resv->iov_base + resv->iov_len)) { in svc_generic_dispatch()
1285 svc_process_common(struct svc_rqst *rqstp, struct kvec *argv, struct kvec *resv) in svc_process_common() argument
1308 svc_putu32(resv, rqstp->rq_xid); in svc_process_common()
1313 svc_putnl(resv, 1); /* REPLY */ in svc_process_common()
1319 reply_statp = resv->iov_base + resv->iov_len; in svc_process_common()
1321 svc_putnl(resv, 0); /* ACCEPT */ in svc_process_common()
1385 statp = resv->iov_base +resv->iov_len; in svc_process_common()
1386 svc_putnl(resv, RPC_SUCCESS); in svc_process_common()
1411 resv->iov_len = ((void*)statp) - resv->iov_base + 4; in svc_process_common()
[all …]
/Linux-v5.15/drivers/gpu/drm/i915/gem/
Di915_gem_wait.c36 i915_gem_object_wait_reservation(struct dma_resv *resv, in i915_gem_object_wait_reservation() argument
48 ret = dma_resv_get_fences(resv, &excl, &count, &shared); in i915_gem_object_wait_reservation()
76 excl = dma_resv_get_excl_unlocked(resv); in i915_gem_object_wait_reservation()
89 dma_resv_prune(resv); in i915_gem_object_wait_reservation()
161 ret = dma_resv_get_fences(obj->base.resv, &excl, &count, in i915_gem_object_wait_priority()
173 excl = dma_resv_get_excl_unlocked(obj->base.resv); in i915_gem_object_wait_priority()
197 timeout = i915_gem_object_wait_reservation(obj->base.resv, in i915_gem_object_wait()
Di915_gem_busy.c113 seq = raw_read_seqcount(&obj->base.resv->seq); in i915_gem_busy_ioctl()
116 args->busy = busy_check_writer(dma_resv_excl_fence(obj->base.resv)); in i915_gem_busy_ioctl()
119 list = dma_resv_shared_list(obj->base.resv); in i915_gem_busy_ioctl()
131 if (args->busy && read_seqcount_retry(&obj->base.resv->seq, seq)) in i915_gem_busy_ioctl()
/Linux-v5.15/mm/
Dhugetlb.c250 get_file_region_entry_from_cache(struct resv_map *resv, long from, long to) in get_file_region_entry_from_cache() argument
254 VM_BUG_ON(resv->region_cache_count <= 0); in get_file_region_entry_from_cache()
256 resv->region_cache_count--; in get_file_region_entry_from_cache()
257 nrg = list_first_entry(&resv->region_cache, struct file_region, link); in get_file_region_entry_from_cache()
280 struct resv_map *resv, in record_hugetlb_cgroup_uncharge_info() argument
299 if (!resv->pages_per_hpage) in record_hugetlb_cgroup_uncharge_info()
300 resv->pages_per_hpage = pages_per_huge_page(h); in record_hugetlb_cgroup_uncharge_info()
304 VM_BUG_ON(resv->pages_per_hpage != pages_per_huge_page(h)); in record_hugetlb_cgroup_uncharge_info()
333 static void coalesce_file_region(struct resv_map *resv, struct file_region *rg) in coalesce_file_region() argument
338 if (&prg->link != &resv->regions && prg->to == rg->from && in coalesce_file_region()
[all …]
/Linux-v5.15/net/sunrpc/auth_gss/
Dsvcauth_gss.c693 svc_safe_putnetobj(struct kvec *resv, struct xdr_netobj *o) in svc_safe_putnetobj() argument
697 if (resv->iov_len + 4 > PAGE_SIZE) in svc_safe_putnetobj()
699 svc_putnl(resv, o->len); in svc_safe_putnetobj()
700 p = resv->iov_base + resv->iov_len; in svc_safe_putnetobj()
701 resv->iov_len += round_up_to_quad(o->len); in svc_safe_putnetobj()
702 if (resv->iov_len > PAGE_SIZE) in svc_safe_putnetobj()
1209 gss_write_resv(struct kvec *resv, size_t size_limit, in gss_write_resv() argument
1213 if (resv->iov_len + 4 > size_limit) in gss_write_resv()
1215 svc_putnl(resv, RPC_SUCCESS); in gss_write_resv()
1216 if (svc_safe_putnetobj(resv, out_handle)) in gss_write_resv()
[all …]
/Linux-v5.15/drivers/infiniband/core/
Dumem_dmabuf.c21 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_map_pages()
68 fence = dma_resv_excl_fence(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_map_pages()
78 dma_resv_assert_held(umem_dmabuf->attach->dmabuf->resv); in ib_umem_dmabuf_unmap_pages()
170 dma_resv_lock(dmabuf->resv, NULL); in ib_umem_dmabuf_release()
172 dma_resv_unlock(dmabuf->resv); in ib_umem_dmabuf_release()
/Linux-v5.15/drivers/net/ethernet/netronome/nfp/crypto/
Dfw.h17 u8 resv[2]; member
37 u8 resv[3]; member
86 u8 resv[3]; member

123456789