Searched refs:s_fence (Results 1 – 11 of 11) sorted by relevance
242 drm_sched_fence_finished(job->s_fence); in drm_sched_entity_kill_jobs_cb()243 WARN_ON(job->s_fence->parent); in drm_sched_entity_kill_jobs_cb()244 dma_fence_put(&job->s_fence->finished); in drm_sched_entity_kill_jobs_cb()327 struct drm_sched_fence *s_fence = job->s_fence; in drm_sched_entity_fini() local328 drm_sched_fence_scheduled(s_fence); in drm_sched_entity_fini()329 dma_fence_set_error(&s_fence->finished, -ESRCH); in drm_sched_entity_fini()422 struct drm_sched_fence *s_fence; in drm_sched_dependency_optimized() local428 s_fence = to_drm_sched_fence(fence); in drm_sched_dependency_optimized()429 if (s_fence && s_fence->sched == sched) in drm_sched_dependency_optimized()440 struct drm_sched_fence *s_fence; in drm_sched_entity_add_dependency_cb() local[all …]
52 __entry->fence = &sched_job->s_fence->finished;
68 struct drm_sched_fence *s_fence = to_drm_sched_fence(f); in amdgpu_sync_same_dev() local70 if (s_fence) { in amdgpu_sync_same_dev()73 ring = container_of(s_fence->sched, struct amdgpu_ring, sched); in amdgpu_sync_same_dev()89 struct drm_sched_fence *s_fence; in amdgpu_sync_get_owner() local95 s_fence = to_drm_sched_fence(f); in amdgpu_sync_get_owner()96 if (s_fence) in amdgpu_sync_get_owner()97 return s_fence->owner; in amdgpu_sync_get_owner()268 struct drm_sched_fence *s_fence = to_drm_sched_fence(f); in amdgpu_sync_peek_fence() local276 if (ring && s_fence) { in amdgpu_sync_peek_fence()280 if (s_fence->sched == &ring->sched) { in amdgpu_sync_peek_fence()[all …]
98 f = job->base.s_fence ? &job->base.s_fence->finished : job->fence; in amdgpu_job_free_resources()141 *f = dma_fence_get(&job->base.s_fence->finished); in amdgpu_job_submit()189 &job->base.s_fence->finished, in amdgpu_job_dependency()208 finished = &job->base.s_fence->finished; in amdgpu_job_run()
38 job->base.s_fence->finished.ops->get_timeline_name(&job->base.s_fence->finished)179 __entry->context = job->base.s_fence->finished.context;180 __entry->seqno = job->base.s_fence->finished.seqno;204 __entry->context = job->base.s_fence->finished.context;205 __entry->seqno = job->base.s_fence->finished.seqno;
142 fence_ctx = job->base.s_fence->scheduled.context; in amdgpu_ib_schedule()
1232 p->fence = dma_fence_get(&job->base.s_fence->finished); in amdgpu_cs_submit()1237 dma_fence_put(&job->base.s_fence->finished); in amdgpu_cs_submit()1270 dma_fence_put(&job->base.s_fence->finished); in amdgpu_cs_submit()1271 job->base.s_fence = NULL; in amdgpu_cs_submit()
77 if (likely(!sched_job->s_fence->finished.error)) in etnaviv_sched_run_job()157 submit->out_fence = dma_fence_get(&submit->sched_job.s_fence->finished); in etnaviv_sched_push_job()
212 struct dma_fence *out_fence = &exec->render.base.s_fence->finished; in v3d_attach_object_fences()562 dma_fence_get(&exec->bin.base.s_fence->finished); in v3d_submit_cl_ioctl()588 &exec->render.base.s_fence->finished); in v3d_submit_cl_ioctl()
89 if (unlikely(job->base.s_fence->finished.error)) in v3d_job_run()
187 struct drm_sched_fence *s_fence; member