/Linux-v4.19/fs/ocfs2/ |
D | reservations.c | 53 struct ocfs2_alloc_reservation *resv) in ocfs2_resv_window_bits() argument 58 if (!(resv->r_flags & OCFS2_RESV_FLAG_DIR)) { in ocfs2_resv_window_bits() 67 static inline unsigned int ocfs2_resv_end(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_end() argument 69 if (resv->r_len) in ocfs2_resv_end() 70 return resv->r_start + resv->r_len - 1; in ocfs2_resv_end() 71 return resv->r_start; in ocfs2_resv_end() 74 static inline int ocfs2_resv_empty(struct ocfs2_alloc_reservation *resv) in ocfs2_resv_empty() argument 76 return !!(resv->r_len == 0); in ocfs2_resv_empty() 90 struct ocfs2_alloc_reservation *resv; in ocfs2_dump_resv() local 98 resv = rb_entry(node, struct ocfs2_alloc_reservation, r_node); in ocfs2_dump_resv() [all …]
|
D | reservations.h | 64 void ocfs2_resv_init_once(struct ocfs2_alloc_reservation *resv); 67 void ocfs2_resv_set_type(struct ocfs2_alloc_reservation *resv, 81 struct ocfs2_alloc_reservation *resv); 137 struct ocfs2_alloc_reservation *resv, 156 struct ocfs2_alloc_reservation *resv,
|
/Linux-v4.19/fs/xfs/libxfs/ |
D | xfs_ag_resv.c | 139 struct xfs_ag_resv *resv; in __xfs_ag_resv_free() local 145 resv = xfs_perag_resv(pag, type); in __xfs_ag_resv_free() 147 pag->pag_mount->m_ag_max_usable += resv->ar_asked; in __xfs_ag_resv_free() 154 oldresv = resv->ar_orig_reserved; in __xfs_ag_resv_free() 156 oldresv = resv->ar_reserved; in __xfs_ag_resv_free() 158 resv->ar_reserved = 0; in __xfs_ag_resv_free() 159 resv->ar_asked = 0; in __xfs_ag_resv_free() 160 resv->ar_orig_reserved = 0; in __xfs_ag_resv_free() 191 struct xfs_ag_resv *resv; in __xfs_ag_resv_init() local 239 resv = xfs_perag_resv(pag, type); in __xfs_ag_resv_init() [all …]
|
/Linux-v4.19/drivers/gpu/drm/i915/ |
D | i915_gem_batch_pool.c | 99 struct reservation_object *resv = obj->resv; in i915_gem_batch_pool_get() local 101 if (!reservation_object_test_signaled_rcu(resv, true)) in i915_gem_batch_pool_get() 115 if (rcu_access_pointer(resv->fence)) { in i915_gem_batch_pool_get() 116 reservation_object_lock(resv, NULL); in i915_gem_batch_pool_get() 117 reservation_object_add_excl_fence(resv, NULL); in i915_gem_batch_pool_get() 118 reservation_object_unlock(resv); in i915_gem_batch_pool_get() 122 GEM_BUG_ON(!reservation_object_test_signaled_rcu(obj->resv, in i915_gem_batch_pool_get()
|
D | i915_gem_clflush.c | 168 obj->resv, NULL, in i915_gem_clflush_object() 172 reservation_object_lock(obj->resv, NULL); in i915_gem_clflush_object() 173 reservation_object_add_excl_fence(obj->resv, &clflush->dma); in i915_gem_clflush_object() 174 reservation_object_unlock(obj->resv); in i915_gem_clflush_object()
|
/Linux-v4.19/drivers/gpu/drm/ttm/ |
D | ttm_execbuf_util.c | 42 reservation_object_unlock(bo->resv); in ttm_eu_backoff_reservation_reverse() 73 reservation_object_unlock(bo->resv); in ttm_eu_backoff_reservation() 116 reservation_object_unlock(bo->resv); in ttm_eu_reserve_buffers() 132 ret = reservation_object_reserve_shared(bo->resv); in ttm_eu_reserve_buffers() 145 ret = ww_mutex_lock_slow_interruptible(&bo->resv->lock, in ttm_eu_reserve_buffers() 148 ww_mutex_lock_slow(&bo->resv->lock, ticket); in ttm_eu_reserve_buffers() 154 ret = reservation_object_reserve_shared(bo->resv); in ttm_eu_reserve_buffers() 205 reservation_object_add_shared_fence(bo->resv, fence); in ttm_eu_fence_buffer_objects() 207 reservation_object_add_excl_fence(bo->resv, fence); in ttm_eu_fence_buffer_objects() 209 reservation_object_unlock(bo->resv); in ttm_eu_fence_buffer_objects()
|
D | ttm_bo.c | 166 reservation_object_assert_held(bo->resv); in ttm_bo_add_to_lru() 219 reservation_object_assert_held(bo->resv); in ttm_bo_move_to_lru_tail() 349 if (bo->resv == &bo->ttm_resv) in ttm_bo_individualize_resv() 354 r = reservation_object_copy_fences(&bo->ttm_resv, bo->resv); in ttm_bo_individualize_resv() 374 reservation_object_held(bo->resv)); in ttm_bo_flush_all_fences() 392 reservation_object_wait_timeout_rcu(bo->resv, true, false, in ttm_bo_cleanup_refs_or_queue() 399 ret = reservation_object_trylock(bo->resv) ? 0 : -EBUSY; in ttm_bo_cleanup_refs_or_queue() 404 if (bo->resv != &bo->ttm_resv) in ttm_bo_cleanup_refs_or_queue() 408 reservation_object_unlock(bo->resv); in ttm_bo_cleanup_refs_or_queue() 424 reservation_object_unlock(bo->resv); in ttm_bo_cleanup_refs_or_queue() [all …]
|
/Linux-v4.19/drivers/gpu/drm/vgem/ |
D | vgem_fence.c | 158 struct reservation_object *resv; in vgem_fence_attach_ioctl() local 184 resv = obj->dma_buf->resv; in vgem_fence_attach_ioctl() 185 if (!reservation_object_test_signaled_rcu(resv, in vgem_fence_attach_ioctl() 193 reservation_object_lock(resv, NULL); in vgem_fence_attach_ioctl() 195 reservation_object_add_excl_fence(resv, fence); in vgem_fence_attach_ioctl() 196 else if ((ret = reservation_object_reserve_shared(resv)) == 0) in vgem_fence_attach_ioctl() 197 reservation_object_add_shared_fence(resv, fence); in vgem_fence_attach_ioctl() 198 reservation_object_unlock(resv); in vgem_fence_attach_ioctl()
|
/Linux-v4.19/drivers/dma-buf/ |
D | dma-buf.c | 76 if (dmabuf->resv == (struct reservation_object *)&dmabuf[1]) in dma_buf_release() 77 reservation_object_fini(dmabuf->resv); in dma_buf_release() 163 struct reservation_object *resv; in dma_buf_poll() local 170 if (!dmabuf || !dmabuf->resv) in dma_buf_poll() 173 resv = dmabuf->resv; in dma_buf_poll() 182 seq = read_seqcount_begin(&resv->seq); in dma_buf_poll() 185 fobj = rcu_dereference(resv->fence); in dma_buf_poll() 190 fence_excl = rcu_dereference(resv->fence_excl); in dma_buf_poll() 191 if (read_seqcount_retry(&resv->seq, seq)) { in dma_buf_poll() 392 struct reservation_object *resv = exp_info->resv; in dma_buf_export() local [all …]
|
/Linux-v4.19/drivers/gpu/drm/radeon/ |
D | radeon_prime.c | 64 struct reservation_object *resv = attach->dmabuf->resv; in radeon_gem_prime_import_sg_table() local 69 ww_mutex_lock(&resv->lock, NULL); in radeon_gem_prime_import_sg_table() 71 RADEON_GEM_DOMAIN_GTT, 0, sg, resv, &bo); in radeon_gem_prime_import_sg_table() 72 ww_mutex_unlock(&resv->lock); in radeon_gem_prime_import_sg_table() 122 return bo->tbo.resv; in radeon_gem_prime_res_obj()
|
D | radeon_benchmark.c | 38 struct reservation_object *resv) in radeon_benchmark_do_move() argument 51 resv); in radeon_benchmark_do_move() 56 resv); in radeon_benchmark_do_move() 125 dobj->tbo.resv); in radeon_benchmark_move() 136 dobj->tbo.resv); in radeon_benchmark_move()
|
D | radeon_sync.c | 91 struct reservation_object *resv, in radeon_sync_resv() argument 101 f = reservation_object_get_excl(resv); in radeon_sync_resv() 108 flist = reservation_object_get_list(resv); in radeon_sync_resv() 114 reservation_object_held(resv)); in radeon_sync_resv()
|
/Linux-v4.19/net/sunrpc/ |
D | svc.c | 1151 svc_process_common(struct svc_rqst *rqstp, struct kvec *argv, struct kvec *resv) in svc_process_common() argument 1177 svc_putu32(resv, rqstp->rq_xid); in svc_process_common() 1182 svc_putnl(resv, 1); /* REPLY */ in svc_process_common() 1188 reply_statp = resv->iov_base + resv->iov_len; in svc_process_common() 1190 svc_putnl(resv, 0); /* ACCEPT */ in svc_process_common() 1261 statp = resv->iov_base +resv->iov_len; in svc_process_common() 1262 svc_putnl(resv, RPC_SUCCESS); in svc_process_common() 1302 !procp->pc_encode(rqstp, resv->iov_base + resv->iov_len)) { in svc_process_common() 1319 resv->iov_len = ((void*)statp) - resv->iov_base + 4; in svc_process_common() 1351 svc_putnl(resv, 1); /* REJECT */ in svc_process_common() [all …]
|
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_object.c | 263 bp.resv = NULL; in amdgpu_bo_create_reserved() 430 .resv = bp->resv, in amdgpu_bo_do_create() 499 NULL, bp->resv, &amdgpu_bo_destroy); in amdgpu_bo_do_create() 515 r = amdgpu_fill_buffer(bo, 0, bo->tbo.resv, &fence); in amdgpu_bo_do_create() 524 if (!bp->resv) in amdgpu_bo_do_create() 537 if (!bp->resv) in amdgpu_bo_do_create() 538 ww_mutex_unlock(&bo->tbo.resv->lock); in amdgpu_bo_do_create() 560 bp.resv = bo->tbo.resv; in amdgpu_bo_create_shadow() 600 if (!bp->resv) in amdgpu_bo_create() 601 WARN_ON(reservation_object_lock((*bo_ptr)->tbo.resv, in amdgpu_bo_create() [all …]
|
D | amdgpu_prime.c | 158 struct reservation_object *resv = attach->dmabuf->resv; in amdgpu_gem_prime_import_sg_table() local 170 bp.resv = resv; in amdgpu_gem_prime_import_sg_table() 171 ww_mutex_lock(&resv->lock, NULL); in amdgpu_gem_prime_import_sg_table() 183 ww_mutex_unlock(&resv->lock); in amdgpu_gem_prime_import_sg_table() 187 ww_mutex_unlock(&resv->lock); in amdgpu_gem_prime_import_sg_table() 225 r = reservation_object_wait_timeout_rcu(bo->tbo.resv, in amdgpu_gem_map_attach() 291 return bo->tbo.resv; in amdgpu_gem_prime_res_obj()
|
D | amdgpu_sync.c | 194 struct reservation_object *resv, in amdgpu_sync_resv() argument 203 if (resv == NULL) in amdgpu_sync_resv() 207 f = reservation_object_get_excl(resv); in amdgpu_sync_resv() 210 flist = reservation_object_get_list(resv); in amdgpu_sync_resv() 216 reservation_object_held(resv)); in amdgpu_sync_resv()
|
D | amdgpu_vm.c | 162 if (bo->tbo.resv != vm->root.base.bo->tbo.resv) in amdgpu_vm_bo_base_init() 392 r = reservation_object_reserve_shared(bo->tbo.resv); in amdgpu_vm_clear_bo() 424 r = amdgpu_sync_resv(adev, &job->sync, bo->tbo.resv, in amdgpu_vm_clear_bo() 508 struct reservation_object *resv = vm->root.base.bo->tbo.resv; in amdgpu_vm_alloc_levels() local 521 bp.resv = resv; in amdgpu_vm_alloc_levels() 953 amdgpu_sync_resv(adev, &sync, vm->root.base.bo->tbo.resv, owner, false); in amdgpu_vm_wait_pd() 1122 amdgpu_sync_resv(adev, &job->sync, root->tbo.resv, in amdgpu_vm_update_directories() 1475 r = amdgpu_sync_resv(adev, &job->sync, vm->root.base.bo->tbo.resv, in amdgpu_vm_bo_update_mapping() 1480 r = reservation_object_reserve_shared(vm->root.base.bo->tbo.resv); in amdgpu_vm_bo_update_mapping() 1664 exclusive = reservation_object_get_excl(bo->tbo.resv); in amdgpu_vm_bo_update() [all …]
|
D | amdgpu_ttm.h | 88 struct reservation_object *resv, 95 struct reservation_object *resv, 99 struct reservation_object *resv,
|
/Linux-v4.19/drivers/gpu/drm/mediatek/ |
D | mtk_drm_fb.c | 67 struct reservation_object *resv; in mtk_fb_wait() local 74 if (!gem || !gem->dma_buf || !gem->dma_buf->resv) in mtk_fb_wait() 77 resv = gem->dma_buf->resv; in mtk_fb_wait() 78 ret = reservation_object_wait_timeout_rcu(resv, false, true, in mtk_fb_wait()
|
/Linux-v4.19/net/sunrpc/auth_gss/ |
D | svcauth_gss.c | 638 svc_safe_putnetobj(struct kvec *resv, struct xdr_netobj *o) in svc_safe_putnetobj() argument 642 if (resv->iov_len + 4 > PAGE_SIZE) in svc_safe_putnetobj() 644 svc_putnl(resv, o->len); in svc_safe_putnetobj() 645 p = resv->iov_base + resv->iov_len; in svc_safe_putnetobj() 646 resv->iov_len += round_up_to_quad(o->len); in svc_safe_putnetobj() 647 if (resv->iov_len > PAGE_SIZE) in svc_safe_putnetobj() 1092 gss_write_resv(struct kvec *resv, size_t size_limit, in gss_write_resv() argument 1096 if (resv->iov_len + 4 > size_limit) in gss_write_resv() 1098 svc_putnl(resv, RPC_SUCCESS); in gss_write_resv() 1099 if (svc_safe_putnetobj(resv, out_handle)) in gss_write_resv() [all …]
|
/Linux-v4.19/mm/ |
D | hugetlb.c | 259 static long region_add(struct resv_map *resv, long f, long t) in region_add() argument 261 struct list_head *head = &resv->regions; in region_add() 265 spin_lock(&resv->lock); in region_add() 278 VM_BUG_ON(resv->region_cache_count <= 0); in region_add() 280 resv->region_cache_count--; in region_add() 281 nrg = list_first_entry(&resv->region_cache, struct file_region, in region_add() 327 resv->adds_in_progress--; in region_add() 328 spin_unlock(&resv->lock); in region_add() 355 static long region_chg(struct resv_map *resv, long f, long t) in region_chg() argument 357 struct list_head *head = &resv->regions; in region_chg() [all …]
|
/Linux-v4.19/include/drm/ttm/ |
D | ttm_bo_driver.h | 634 success = reservation_object_trylock(bo->resv); in __ttm_bo_reserve() 639 ret = reservation_object_lock_interruptible(bo->resv, ticket); in __ttm_bo_reserve() 641 ret = reservation_object_lock(bo->resv, ticket); in __ttm_bo_reserve() 725 ret = ww_mutex_lock_slow_interruptible(&bo->resv->lock, in ttm_bo_reserve_slowpath() 728 ww_mutex_lock_slow(&bo->resv->lock, ticket); in ttm_bo_reserve_slowpath() 752 reservation_object_unlock(bo->resv); in ttm_bo_unreserve()
|
D | ttm_bo_api.h | 231 struct reservation_object *resv; member 276 struct reservation_object *resv; member 527 struct reservation_object *resv, 571 struct sg_table *sg, struct reservation_object *resv,
|
/Linux-v4.19/drivers/gpu/drm/msm/ |
D | msm_gem.c | 626 fobj = reservation_object_get_list(msm_obj->resv); in msm_gem_sync_object() 628 fence = reservation_object_get_excl(msm_obj->resv); in msm_gem_sync_object() 642 reservation_object_held(msm_obj->resv)); in msm_gem_sync_object() 660 reservation_object_add_excl_fence(msm_obj->resv, fence); in msm_gem_move_to_active() 662 reservation_object_add_shared_fence(msm_obj->resv, fence); in msm_gem_move_to_active() 688 ret = reservation_object_wait_timeout_rcu(msm_obj->resv, write, in msm_gem_cpu_prep() 720 struct reservation_object *robj = msm_obj->resv; in msm_gem_describe() 823 if (msm_obj->resv == &msm_obj->_resv) in msm_gem_free_object() 824 reservation_object_fini(msm_obj->resv); in msm_gem_free_object() 854 struct reservation_object *resv, in msm_gem_new_impl() argument [all …]
|
/Linux-v4.19/kernel/irq/ |
D | affinity.c | 259 int resv = affd->pre_vectors + affd->post_vectors; in irq_calc_affinity_vectors() local 260 int vecs = maxvec - resv; in irq_calc_affinity_vectors() 263 if (resv > minvec) in irq_calc_affinity_vectors() 267 ret = min_t(int, cpumask_weight(cpu_possible_mask), vecs) + resv; in irq_calc_affinity_vectors()
|