Lines Matching refs:resv
319 if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv) in amdgpu_vm_bo_base_init()
322 dma_resv_assert_held(vm->root.bo->tbo.base.resv); in amdgpu_vm_bo_base_init()
864 struct dma_resv *resv, uint64_t start, uint64_t last, in amdgpu_vm_update_range() argument
924 r = vm->update_funcs->prepare(¶ms, resv, sync_mode); in amdgpu_vm_update_range()
1019 if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv && in amdgpu_vm_bo_get_memory()
1020 !dma_resv_trylock(bo->tbo.base.resv)) in amdgpu_vm_bo_get_memory()
1024 if (bo->tbo.base.resv != vm->root.bo->tbo.base.resv) in amdgpu_vm_bo_get_memory()
1025 dma_resv_unlock(bo->tbo.base.resv); in amdgpu_vm_bo_get_memory()
1076 struct dma_resv *resv; in amdgpu_vm_bo_update() local
1083 resv = vm->root.bo->tbo.base.resv; in amdgpu_vm_bo_update()
1087 resv = bo->tbo.base.resv; in amdgpu_vm_bo_update()
1118 if (clear || (bo && bo->tbo.base.resv == in amdgpu_vm_bo_update()
1119 vm->root.bo->tbo.base.resv)) in amdgpu_vm_bo_update()
1149 resv, mapping->start, mapping->last, in amdgpu_vm_bo_update()
1161 if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv) { in amdgpu_vm_bo_update()
1299 struct dma_resv *resv = vm->root.bo->tbo.base.resv; in amdgpu_vm_prt_fini() local
1303 dma_resv_for_each_fence(&cursor, resv, DMA_RESV_USAGE_BOOKKEEP, fence) { in amdgpu_vm_prt_fini()
1329 struct dma_resv *resv = vm->root.bo->tbo.base.resv; in amdgpu_vm_clear_freed() local
1344 r = amdgpu_vm_update_range(adev, vm, false, false, true, resv, in amdgpu_vm_clear_freed()
1383 struct dma_resv *resv; in amdgpu_vm_handle_moved() local
1403 resv = bo_va->base.bo->tbo.base.resv; in amdgpu_vm_handle_moved()
1407 if (!amdgpu_vm_debug && dma_resv_trylock(resv)) in amdgpu_vm_handle_moved()
1418 dma_resv_unlock(resv); in amdgpu_vm_handle_moved()
1461 dma_resv_assert_held(bo->tbo.base.resv); in amdgpu_vm_bo_add()
1495 if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv && in amdgpu_vm_bo_insert_map()
1777 if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv && in amdgpu_vm_bo_clear_mappings()
1792 if (bo && bo->tbo.base.resv == vm->root.bo->tbo.base.resv && in amdgpu_vm_bo_clear_mappings()
1841 if (dma_resv_locking_ctx(bo->tbo.base.resv) != in amdgpu_vm_bo_trace_cs()
1868 dma_resv_assert_held(vm->root.bo->tbo.base.resv); in amdgpu_vm_bo_del()
1871 dma_resv_assert_held(bo->tbo.base.resv); in amdgpu_vm_bo_del()
1872 if (bo->tbo.base.resv == vm->root.bo->tbo.base.resv) in amdgpu_vm_bo_del()
1927 if (!dma_resv_test_signaled(bo->tbo.base.resv, DMA_RESV_USAGE_BOOKKEEP)) in amdgpu_vm_evictable()
1966 if (evicted && bo->tbo.base.resv == vm->root.bo->tbo.base.resv) { in amdgpu_vm_bo_invalidate()
1977 else if (bo->tbo.base.resv == vm->root.bo->tbo.base.resv) in amdgpu_vm_bo_invalidate()
2107 timeout = dma_resv_wait_timeout(vm->root.bo->tbo.base.resv, in amdgpu_vm_wait_idle()
2186 r = dma_resv_reserve_fences(root_bo->tbo.base.resv, 1); in amdgpu_vm_init()
2615 r = dma_resv_reserve_fences(root->tbo.base.resv, 1); in amdgpu_vm_handle_fault()