Lines Matching refs:e

135 	struct amdgpu_sync_entry *e;  in amdgpu_sync_add_later()  local
137 hash_for_each_possible(sync->fences, e, node, f->context) { in amdgpu_sync_add_later()
138 if (unlikely(e->fence->context != f->context)) in amdgpu_sync_add_later()
141 amdgpu_sync_keep_later(&e->fence, f); in amdgpu_sync_add_later()
144 e->explicit |= explicit; in amdgpu_sync_add_later()
161 struct amdgpu_sync_entry *e; in amdgpu_sync_fence() local
172 e = kmem_cache_alloc(amdgpu_sync_slab, GFP_KERNEL); in amdgpu_sync_fence()
173 if (!e) in amdgpu_sync_fence()
176 e->explicit = explicit; in amdgpu_sync_fence()
178 hash_add(sync->fences, &e->node, f->context); in amdgpu_sync_fence()
179 e->fence = dma_fence_get(f); in amdgpu_sync_fence()
262 struct amdgpu_sync_entry *e; in amdgpu_sync_peek_fence() local
266 hash_for_each_safe(sync->fences, i, tmp, e, node) { in amdgpu_sync_peek_fence()
267 struct dma_fence *f = e->fence; in amdgpu_sync_peek_fence()
271 hash_del(&e->node); in amdgpu_sync_peek_fence()
273 kmem_cache_free(amdgpu_sync_slab, e); in amdgpu_sync_peek_fence()
304 struct amdgpu_sync_entry *e; in amdgpu_sync_get_fence() local
308 hash_for_each_safe(sync->fences, i, tmp, e, node) { in amdgpu_sync_get_fence()
310 f = e->fence; in amdgpu_sync_get_fence()
312 *explicit = e->explicit; in amdgpu_sync_get_fence()
314 hash_del(&e->node); in amdgpu_sync_get_fence()
315 kmem_cache_free(amdgpu_sync_slab, e); in amdgpu_sync_get_fence()
336 struct amdgpu_sync_entry *e; in amdgpu_sync_clone() local
341 hash_for_each_safe(source->fences, i, tmp, e, node) { in amdgpu_sync_clone()
342 f = e->fence; in amdgpu_sync_clone()
344 r = amdgpu_sync_fence(NULL, clone, f, e->explicit); in amdgpu_sync_clone()
348 hash_del(&e->node); in amdgpu_sync_clone()
350 kmem_cache_free(amdgpu_sync_slab, e); in amdgpu_sync_clone()
362 struct amdgpu_sync_entry *e; in amdgpu_sync_wait() local
366 hash_for_each_safe(sync->fences, i, tmp, e, node) { in amdgpu_sync_wait()
367 r = dma_fence_wait(e->fence, intr); in amdgpu_sync_wait()
371 hash_del(&e->node); in amdgpu_sync_wait()
372 dma_fence_put(e->fence); in amdgpu_sync_wait()
373 kmem_cache_free(amdgpu_sync_slab, e); in amdgpu_sync_wait()
388 struct amdgpu_sync_entry *e; in amdgpu_sync_free() local
392 hash_for_each_safe(sync->fences, i, tmp, e, node) { in amdgpu_sync_free()
393 hash_del(&e->node); in amdgpu_sync_free()
394 dma_fence_put(e->fence); in amdgpu_sync_free()
395 kmem_cache_free(amdgpu_sync_slab, e); in amdgpu_sync_free()