Lines Matching refs:sync

50 void amdgpu_sync_create(struct amdgpu_sync *sync)  in amdgpu_sync_create()  argument
52 hash_init(sync->fences); in amdgpu_sync_create()
53 sync->last_vm_update = NULL; in amdgpu_sync_create()
132 static bool amdgpu_sync_add_later(struct amdgpu_sync *sync, struct dma_fence *f, bool explicit) in amdgpu_sync_add_later() argument
136 hash_for_each_possible(sync->fences, e, node, f->context) { in amdgpu_sync_add_later()
157 int amdgpu_sync_fence(struct amdgpu_device *adev, struct amdgpu_sync *sync, in amdgpu_sync_fence() argument
166 amdgpu_sync_keep_later(&sync->last_vm_update, f); in amdgpu_sync_fence()
168 if (amdgpu_sync_add_later(sync, f, explicit)) in amdgpu_sync_fence()
177 hash_add(sync->fences, &e->node, f->context); in amdgpu_sync_fence()
192 struct amdgpu_sync *sync, in amdgpu_sync_resv() argument
207 r = amdgpu_sync_fence(adev, sync, f, false); in amdgpu_sync_resv()
242 r = amdgpu_sync_fence(adev, sync, f, false); in amdgpu_sync_resv()
258 struct dma_fence *amdgpu_sync_peek_fence(struct amdgpu_sync *sync, in amdgpu_sync_peek_fence() argument
265 hash_for_each_safe(sync->fences, i, tmp, e, node) { in amdgpu_sync_peek_fence()
301 struct dma_fence *amdgpu_sync_get_fence(struct amdgpu_sync *sync, bool *explicit) in amdgpu_sync_get_fence() argument
307 hash_for_each_safe(sync->fences, i, tmp, e, node) { in amdgpu_sync_get_fence()
359 int amdgpu_sync_wait(struct amdgpu_sync *sync, bool intr) in amdgpu_sync_wait() argument
365 hash_for_each_safe(sync->fences, i, tmp, e, node) { in amdgpu_sync_wait()
385 void amdgpu_sync_free(struct amdgpu_sync *sync) in amdgpu_sync_free() argument
391 hash_for_each_safe(sync->fences, i, tmp, e, node) { in amdgpu_sync_free()
397 dma_fence_put(sync->last_vm_update); in amdgpu_sync_free()