/Linux-v5.4/fs/ocfs2/ |
D | reservations.c | 45 struct ocfs2_alloc_reservation *resv) in ocfs2_resv_window_bits() argument 50 if (!(resv->r_flags & OCFS2_RESV_FLAG_DIR)) { in ocfs2_resv_window_bits() 59 static inline unsigned int ocfs2_resv_end(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_end() argument 61 if (resv->r_len) in ocfs2_resv_end() 62 return resv->r_start + resv->r_len - 1; in ocfs2_resv_end() 63 return resv->r_start; in ocfs2_resv_end() 66 static inline int ocfs2_resv_empty(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_empty() argument 68 return !!(resv->r_len == 0); in ocfs2_resv_empty() 82 struct ocfs2_alloc_reservation *resv; in ocfs2_dump_resv() local 90 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_dump_resv() [all …]
|
D | reservations.h | 56 void ocfs2_resv_init_once(struct ocfs2_alloc_reservation *resv); 59 void ocfs2_resv_set_type(struct ocfs2_alloc_reservation *resv, 73 struct ocfs2_alloc_reservation *resv); 129 struct ocfs2_alloc_reservation *resv, 148 struct ocfs2_alloc_reservation *resv,
|
/Linux-v5.4/fs/xfs/libxfs/ |
D | xfs_ag_resv.c | 131 struct xfs_ag_resv *resv; in __xfs_ag_resv_free() local 137 resv = xfs_perag_resv(pag, type); in __xfs_ag_resv_free() 139 pag->pag_mount->m_ag_max_usable += resv->ar_asked; in __xfs_ag_resv_free() 146 oldresv = resv->ar_orig_reserved; in __xfs_ag_resv_free() 148 oldresv = resv->ar_reserved; in __xfs_ag_resv_free() 150 resv->ar_reserved = 0; in __xfs_ag_resv_free() 151 resv->ar_asked = 0; in __xfs_ag_resv_free() 152 resv->ar_orig_reserved = 0; in __xfs_ag_resv_free() 183 struct xfs_ag_resv *resv; in __xfs_ag_resv_init() local 231 resv = xfs_perag_resv(pag, type); in __xfs_ag_resv_init() [all …]
|
/Linux-v5.4/drivers/gpu/drm/ttm/ |
D | ttm_execbuf_util.c | 42 dma_resv_unlock(bo->base.resv); in ttm_eu_backoff_reservation_reverse() 74 dma_resv_unlock(bo->base.resv); in ttm_eu_backoff_reservation() 117 dma_resv_unlock(bo->base.resv); in ttm_eu_reserve_buffers() 133 ret = dma_resv_reserve_shared(bo->base.resv, in ttm_eu_reserve_buffers() 147 ret = dma_resv_lock_slow_interruptible(bo->base.resv, in ttm_eu_reserve_buffers() 150 dma_resv_lock_slow(bo->base.resv, ticket); in ttm_eu_reserve_buffers() 156 ret = dma_resv_reserve_shared(bo->base.resv, in ttm_eu_reserve_buffers() 204 dma_resv_add_shared_fence(bo->base.resv, fence); in ttm_eu_fence_buffer_objects() 206 dma_resv_add_excl_fence(bo->base.resv, fence); in ttm_eu_fence_buffer_objects() 211 dma_resv_unlock(bo->base.resv); in ttm_eu_fence_buffer_objects()
|
D | ttm_bo.c | 176 dma_resv_assert_held(bo->base.resv); in ttm_bo_add_mem_to_lru() 248 dma_resv_assert_held(bo->base.resv); in ttm_bo_move_to_lru_tail() 281 dma_resv_assert_held(pos->first->base.resv); in ttm_bo_bulk_move_lru_tail() 282 dma_resv_assert_held(pos->last->base.resv); in ttm_bo_bulk_move_lru_tail() 296 dma_resv_assert_held(pos->first->base.resv); in ttm_bo_bulk_move_lru_tail() 297 dma_resv_assert_held(pos->last->base.resv); in ttm_bo_bulk_move_lru_tail() 311 dma_resv_assert_held(pos->first->base.resv); in ttm_bo_bulk_move_lru_tail() 312 dma_resv_assert_held(pos->last->base.resv); in ttm_bo_bulk_move_lru_tail() 443 if (bo->base.resv == &bo->base._resv) in ttm_bo_individualize_resv() 448 r = dma_resv_copy_fences(&bo->base._resv, bo->base.resv); in ttm_bo_individualize_resv() [all …]
|
/Linux-v5.4/drivers/gpu/drm/i915/gem/ |
D | i915_gem_wait.c | 34 i915_gem_object_wait_reservation(struct dma_resv *resv, in i915_gem_object_wait_reservation() argument 46 ret = dma_resv_get_fences_rcu(resv, in i915_gem_object_wait_reservation() 75 excl = dma_resv_get_excl_rcu(resv); in i915_gem_object_wait_reservation() 87 if (prune_fences && dma_resv_trylock(resv)) { in i915_gem_object_wait_reservation() 88 if (dma_resv_test_signaled_rcu(resv, true)) in i915_gem_object_wait_reservation() 89 dma_resv_add_excl_fence(resv, NULL); in i915_gem_object_wait_reservation() 90 dma_resv_unlock(resv); in i915_gem_object_wait_reservation() 143 ret = dma_resv_get_fences_rcu(obj->base.resv, in i915_gem_object_wait_priority() 155 excl = dma_resv_get_excl_rcu(obj->base.resv); in i915_gem_object_wait_priority() 179 timeout = i915_gem_object_wait_reservation(obj->base.resv, in i915_gem_object_wait()
|
D | i915_gem_busy.c | 113 seq = raw_read_seqcount(&obj->base.resv->seq); in i915_gem_busy_ioctl() 117 busy_check_writer(rcu_dereference(obj->base.resv->fence_excl)); in i915_gem_busy_ioctl() 120 list = rcu_dereference(obj->base.resv->fence); in i915_gem_busy_ioctl() 132 if (args->busy && read_seqcount_retry(&obj->base.resv->seq, seq)) in i915_gem_busy_ioctl()
|
/Linux-v5.4/drivers/gpu/drm/vgem/ |
D | vgem_fence.c | 131 struct dma_resv *resv; in vgem_fence_attach_ioctl() local 153 resv = obj->resv; in vgem_fence_attach_ioctl() 154 if (!dma_resv_test_signaled_rcu(resv, in vgem_fence_attach_ioctl() 162 dma_resv_lock(resv, NULL); in vgem_fence_attach_ioctl() 164 dma_resv_add_excl_fence(resv, fence); in vgem_fence_attach_ioctl() 165 else if ((ret = dma_resv_reserve_shared(resv, 1)) == 0) in vgem_fence_attach_ioctl() 166 dma_resv_add_shared_fence(resv, fence); in vgem_fence_attach_ioctl() 167 dma_resv_unlock(resv); in vgem_fence_attach_ioctl()
|
/Linux-v5.4/net/sunrpc/ |
D | svc.c | 1191 struct kvec *resv = &rqstp->rq_res.head[0]; in svc_generic_dispatch() local 1218 !procp->pc_encode(rqstp, resv->iov_base + resv->iov_len)) { in svc_generic_dispatch() 1283 svc_process_common(struct svc_rqst *rqstp, struct kvec *argv, struct kvec *resv) in svc_process_common() argument 1306 svc_putu32(resv, rqstp->rq_xid); in svc_process_common() 1311 svc_putnl(resv, 1); /* REPLY */ in svc_process_common() 1317 reply_statp = resv->iov_base + resv->iov_len; in svc_process_common() 1319 svc_putnl(resv, 0); /* ACCEPT */ in svc_process_common() 1383 statp = resv->iov_base +resv->iov_len; in svc_process_common() 1384 svc_putnl(resv, RPC_SUCCESS); in svc_process_common() 1409 resv->iov_len = ((void*)statp) - resv->iov_base + 4; in svc_process_common() [all …]
|
/Linux-v5.4/drivers/dma-buf/ |
D | dma-buf.c | 107 if (dmabuf->resv == (struct dma_resv *)&dmabuf[1]) in dma_buf_release() 108 dma_resv_fini(dmabuf->resv); in dma_buf_release() 198 struct dma_resv *resv; in dma_buf_poll() local 205 if (!dmabuf || !dmabuf->resv) in dma_buf_poll() 208 resv = dmabuf->resv; in dma_buf_poll() 217 seq = read_seqcount_begin(&resv->seq); in dma_buf_poll() 220 fobj = rcu_dereference(resv->fence); in dma_buf_poll() 225 fence_excl = rcu_dereference(resv->fence_excl); in dma_buf_poll() 226 if (read_seqcount_retry(&resv->seq, seq)) { in dma_buf_poll() 509 struct dma_resv *resv = exp_info->resv; in dma_buf_export() local [all …]
|
/Linux-v5.4/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_object.c | 253 bp.resv = NULL; in amdgpu_bo_create_reserved() 455 .resv = bp->resv, in amdgpu_bo_do_create() 516 NULL, bp->resv, &amdgpu_bo_destroy); in amdgpu_bo_do_create() 532 r = amdgpu_fill_buffer(bo, 0, bo->tbo.base.resv, &fence); in amdgpu_bo_do_create() 541 if (!bp->resv) in amdgpu_bo_do_create() 554 if (!bp->resv) in amdgpu_bo_do_create() 555 dma_resv_unlock(bo->tbo.base.resv); in amdgpu_bo_do_create() 576 bp.resv = bo->tbo.base.resv; in amdgpu_bo_create_shadow() 616 if (!bp->resv) in amdgpu_bo_create() 617 WARN_ON(dma_resv_lock((*bo_ptr)->tbo.base.resv, in amdgpu_bo_create() [all …]
|
D | amdgpu_dma_buf.c | 219 r = __dma_resv_make_exclusive(bo->tbo.base.resv); in amdgpu_dma_buf_map_attach() 370 struct dma_resv *resv = attach->dmabuf->resv; in amdgpu_gem_prime_import_sg_table() local 382 bp.resv = resv; in amdgpu_gem_prime_import_sg_table() 383 dma_resv_lock(resv, NULL); in amdgpu_gem_prime_import_sg_table() 395 dma_resv_unlock(resv); in amdgpu_gem_prime_import_sg_table() 399 dma_resv_unlock(resv); in amdgpu_gem_prime_import_sg_table()
|
D | amdgpu_sync.c | 193 struct dma_resv *resv, in amdgpu_sync_resv() argument 202 if (resv == NULL) in amdgpu_sync_resv() 206 f = dma_resv_get_excl(resv); in amdgpu_sync_resv() 209 flist = dma_resv_get_list(resv); in amdgpu_sync_resv() 215 dma_resv_held(resv)); in amdgpu_sync_resv()
|
D | amdgpu_ttm.h | 88 struct dma_resv *resv, 95 struct dma_resv *resv, 99 struct dma_resv *resv,
|
D | amdgpu_vm.c | 305 if (bo->tbo.base.resv != vm->root.base.bo->tbo.base.resv) in amdgpu_vm_bo_base_init() 586 if (abo->tbo.base.resv == vm->root.base.bo->tbo.base.resv) in amdgpu_vm_del_from_lru_notify() 835 bp->resv = vm->root.base.bo->tbo.base.resv; in amdgpu_vm_bo_param() 1703 exclusive = dma_resv_get_excl(bo->tbo.base.resv); in amdgpu_vm_bo_update() 1713 if (clear || (bo && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv)) in amdgpu_vm_bo_update() 1744 if (bo && bo->tbo.base.resv == vm->root.base.bo->tbo.base.resv) { in amdgpu_vm_bo_update() 1880 struct dma_resv *resv = vm->root.base.bo->tbo.base.resv; in amdgpu_vm_prt_fini() local 1885 r = dma_resv_get_fences_rcu(resv, &excl, in amdgpu_vm_prt_fini() 1891 dma_resv_wait_timeout_rcu(resv, true, false, in amdgpu_vm_prt_fini() 1979 struct dma_resv *resv; in amdgpu_vm_handle_moved() local [all …]
|
/Linux-v5.4/drivers/gpu/drm/radeon/ |
D | radeon_prime.c | 66 struct dma_resv *resv = attach->dmabuf->resv; in radeon_gem_prime_import_sg_table() local 71 dma_resv_lock(resv, NULL); in radeon_gem_prime_import_sg_table() 73 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); in radeon_gem_prime_import_sg_table() 74 dma_resv_unlock(resv); in radeon_gem_prime_import_sg_table()
|
D | radeon_benchmark.c | 38 struct dma_resv *resv) in radeon_benchmark_do_move() argument 51 resv); in radeon_benchmark_do_move() 56 resv); in radeon_benchmark_do_move() 125 dobj->tbo.base.resv); in radeon_benchmark_move() 136 dobj->tbo.base.resv); in radeon_benchmark_move()
|
D | radeon_sync.c | 90 struct dma_resv *resv, in radeon_sync_resv() argument 100 f = dma_resv_get_excl(resv); in radeon_sync_resv() 107 flist = dma_resv_get_list(resv); in radeon_sync_resv() 113 dma_resv_held(resv)); in radeon_sync_resv()
|
/Linux-v5.4/net/sunrpc/auth_gss/ |
D | svcauth_gss.c | 659 svc_safe_putnetobj(struct kvec *resv, struct xdr_netobj *o) in svc_safe_putnetobj() argument 663 if (resv->iov_len + 4 > PAGE_SIZE) in svc_safe_putnetobj() 665 svc_putnl(resv, o->len); in svc_safe_putnetobj() 666 p = resv->iov_base + resv->iov_len; in svc_safe_putnetobj() 667 resv->iov_len += round_up_to_quad(o->len); in svc_safe_putnetobj() 668 if (resv->iov_len > PAGE_SIZE) in svc_safe_putnetobj() 1113 gss_write_resv(struct kvec *resv, size_t size_limit, in gss_write_resv() argument 1117 if (resv->iov_len + 4 > size_limit) in gss_write_resv() 1119 svc_putnl(resv, RPC_SUCCESS); in gss_write_resv() 1120 if (svc_safe_putnetobj(resv, out_handle)) in gss_write_resv() [all …]
|
/Linux-v5.4/drivers/gpu/drm/i915/gt/ |
D | intel_engine_pool.c | 46 struct dma_resv *resv = node->obj->base.resv; in pool_active() local 49 if (dma_resv_trylock(resv)) { in pool_active() 50 dma_resv_add_excl_fence(resv, NULL); in pool_active() 51 dma_resv_unlock(resv); in pool_active()
|
/Linux-v5.4/drivers/gpu/drm/lima/ |
D | lima_gem.c | 139 err = dma_resv_reserve_shared(bo->gem.resv, 1); in lima_gem_sync_bo() 165 ret = ww_mutex_lock_interruptible(&bos[i]->gem.resv->lock, ctx); in lima_gem_lock_bos() 177 ww_mutex_unlock(&bos[i]->gem.resv->lock); in lima_gem_lock_bos() 180 ww_mutex_unlock(&bos[slow_locked]->gem.resv->lock); in lima_gem_lock_bos() 185 &bos[contended]->gem.resv->lock, ctx); in lima_gem_lock_bos() 202 ww_mutex_unlock(&bos[i]->gem.resv->lock); in lima_gem_unlock_bos() 299 dma_resv_add_excl_fence(bos[i]->gem.resv, fence); in lima_gem_submit() 301 dma_resv_add_shared_fence(bos[i]->gem.resv, fence); in lima_gem_submit()
|
/Linux-v5.4/mm/ |
D | hugetlb.c | 261 static long region_add(struct resv_map *resv, long f, long t) in region_add() argument 263 struct list_head *head = &resv->regions; in region_add() 267 spin_lock(&resv->lock); in region_add() 280 VM_BUG_ON(resv->region_cache_count <= 0); in region_add() 282 resv->region_cache_count--; in region_add() 283 nrg = list_first_entry(&resv->region_cache, struct file_region, in region_add() 329 resv->adds_in_progress--; in region_add() 330 spin_unlock(&resv->lock); in region_add() 357 static long region_chg(struct resv_map *resv, long f, long t) in region_chg() argument 359 struct list_head *head = &resv->regions; in region_chg() [all …]
|
/Linux-v5.4/include/drm/ttm/ |
D | ttm_bo_driver.h | 667 success = dma_resv_trylock(bo->base.resv); in __ttm_bo_reserve() 672 ret = dma_resv_lock_interruptible(bo->base.resv, ticket); in __ttm_bo_reserve() 674 ret = dma_resv_lock(bo->base.resv, ticket); in __ttm_bo_reserve() 758 ret = dma_resv_lock_slow_interruptible(bo->base.resv, in ttm_bo_reserve_slowpath() 761 dma_resv_lock_slow(bo->base.resv, ticket); in ttm_bo_reserve_slowpath() 786 dma_resv_unlock(bo->base.resv); in ttm_bo_unreserve()
|
/Linux-v5.4/arch/sparc/include/asm/ |
D | vio.h | 44 u64 resv[6]; member 71 u16 resv; member 79 u64 resv[5]; member 179 u32 resv; member 200 u16 resv; member 242 u32 resv; member
|
/Linux-v5.4/kernel/irq/ |
D | affinity.c | 499 unsigned int resv = affd->pre_vectors + affd->post_vectors; in irq_calc_affinity_vectors() local 502 if (resv > minvec) in irq_calc_affinity_vectors() 506 set_vecs = maxvec - resv; in irq_calc_affinity_vectors() 513 return resv + min(set_vecs, maxvec - resv); in irq_calc_affinity_vectors()
|