/Linux-v6.1/drivers/gpu/drm/v3d/ |
D | v3d_sched.c | 28 to_v3d_job(struct drm_sched_job *sched_job) in to_v3d_job() argument 30 return container_of(sched_job, struct v3d_job, base); in to_v3d_job() 34 to_bin_job(struct drm_sched_job *sched_job) in to_bin_job() argument 36 return container_of(sched_job, struct v3d_bin_job, base.base); in to_bin_job() 40 to_render_job(struct drm_sched_job *sched_job) in to_render_job() argument 42 return container_of(sched_job, struct v3d_render_job, base.base); in to_render_job() 46 to_tfu_job(struct drm_sched_job *sched_job) in to_tfu_job() argument 48 return container_of(sched_job, struct v3d_tfu_job, base.base); in to_tfu_job() 52 to_csd_job(struct drm_sched_job *sched_job) in to_csd_job() argument 54 return container_of(sched_job, struct v3d_csd_job, base.base); in to_csd_job() [all …]
|
/Linux-v6.1/drivers/gpu/drm/scheduler/ |
D | gpu_scheduler_trace.h | 36 TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity), 37 TP_ARGS(sched_job, entity), 41 __string(name, sched_job->sched->name) 49 __entry->id = sched_job->id; 50 __entry->fence = &sched_job->s_fence->finished; 51 __assign_str(name, sched_job->sched->name); 54 &sched_job->sched->hw_rq_count); 63 TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity), 64 TP_ARGS(sched_job, entity) 68 TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity), [all …]
|
D | sched_entity.c | 33 #define to_drm_sched_job(sched_job) \ argument 34 container_of((sched_job), struct drm_sched_job, queue_node) 421 struct drm_sched_job *sched_job; in drm_sched_entity_pop_job() local 423 sched_job = to_drm_sched_job(spsc_queue_peek(&entity->job_queue)); in drm_sched_entity_pop_job() 424 if (!sched_job) in drm_sched_entity_pop_job() 428 drm_sched_job_dependency(sched_job, entity))) { in drm_sched_entity_pop_job() 429 trace_drm_sched_job_wait_dep(sched_job, entity->dependency); in drm_sched_entity_pop_job() 437 dma_fence_set_error(&sched_job->s_fence->finished, -ECANCELED); in drm_sched_entity_pop_job() 441 entity->last_scheduled = dma_fence_get(&sched_job->s_fence->finished); in drm_sched_entity_pop_job() 451 return sched_job; in drm_sched_entity_pop_job() [all …]
|
D | sched_main.c | 62 #define to_drm_sched_job(sched_job) \ argument 63 container_of((sched_job), struct drm_sched_job, queue_node) 927 struct drm_sched_job *sched_job; in drm_sched_main() local 943 sched_job = drm_sched_entity_pop_job(entity); in drm_sched_main() 945 if (!sched_job) { in drm_sched_main() 950 s_fence = sched_job->s_fence; in drm_sched_main() 953 drm_sched_job_begin(sched_job); in drm_sched_main() 955 trace_drm_run_job(sched_job, entity); in drm_sched_main() 956 fence = sched->ops->run_job(sched_job); in drm_sched_main() 965 r = dma_fence_add_callback(fence, &sched_job->cb, in drm_sched_main() [all …]
|
/Linux-v6.1/drivers/gpu/drm/etnaviv/ |
D | etnaviv_sched.c | 20 static struct dma_fence *etnaviv_sched_run_job(struct drm_sched_job *sched_job) in etnaviv_sched_run_job() argument 22 struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); in etnaviv_sched_run_job() 25 if (likely(!sched_job->s_fence->finished.error)) in etnaviv_sched_run_job() 34 *sched_job) in etnaviv_sched_timedout_job() 36 struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); in etnaviv_sched_timedout_job() 42 drm_sched_stop(&gpu->sched, sched_job); in etnaviv_sched_timedout_job() 65 if(sched_job) in etnaviv_sched_timedout_job() 66 drm_sched_increase_karma(sched_job); in etnaviv_sched_timedout_job() 83 static void etnaviv_sched_free_job(struct drm_sched_job *sched_job) in etnaviv_sched_free_job() argument 85 struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); in etnaviv_sched_free_job() [all …]
|
D | etnaviv_sched.h | 14 struct etnaviv_gem_submit *to_etnaviv_submit(struct drm_sched_job *sched_job) in to_etnaviv_submit() argument 16 return container_of(sched_job, struct etnaviv_gem_submit, sched_job); in to_etnaviv_submit()
|
D | etnaviv_gem_submit.c | 189 ret = drm_sched_job_add_implicit_dependencies(&submit->sched_job, in submit_fence_sync() 532 ret = drm_sched_job_init(&submit->sched_job, in etnaviv_ioctl_gem_submit() 556 ret = drm_sched_job_add_dependency(&submit->sched_job, in etnaviv_ioctl_gem_submit() 616 drm_sched_job_cleanup(&submit->sched_job); in etnaviv_ioctl_gem_submit()
|
D | etnaviv_gem.h | 90 struct drm_sched_job sched_job; member
|
/Linux-v6.1/include/drm/ |
D | gpu_scheduler.h | 358 struct dma_fence *(*dependency)(struct drm_sched_job *sched_job, 367 struct dma_fence *(*run_job)(struct drm_sched_job *sched_job); 411 enum drm_gpu_sched_stat (*timedout_job)(struct drm_sched_job *sched_job); 417 void (*free_job)(struct drm_sched_job *sched_job); 525 void drm_sched_entity_push_job(struct drm_sched_job *sched_job);
|
/Linux-v6.1/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_job.c | 245 static struct dma_fence *amdgpu_job_dependency(struct drm_sched_job *sched_job, in amdgpu_job_dependency() argument 249 struct amdgpu_job *job = to_amdgpu_job(sched_job); in amdgpu_job_dependency() 277 static struct dma_fence *amdgpu_job_run(struct drm_sched_job *sched_job) in amdgpu_job_run() argument 279 struct amdgpu_ring *ring = to_amdgpu_ring(sched_job->sched); in amdgpu_job_run() 285 job = to_amdgpu_job(sched_job); in amdgpu_job_run() 313 #define to_drm_sched_job(sched_job) \ argument 314 container_of((sched_job), struct drm_sched_job, queue_node)
|
D | amdgpu_job.h | 39 #define to_amdgpu_job(sched_job) \ argument 40 container_of((sched_job), struct amdgpu_job, base)
|
D | amdgpu_trace.h | 518 TP_PROTO(struct amdgpu_job *sched_job, struct dma_fence *fence), 519 TP_ARGS(sched_job, fence), 521 __string(ring, sched_job->base.sched->name) 529 __assign_str(ring, sched_job->base.sched->name); 530 __entry->id = sched_job->base.id;
|
/Linux-v6.1/drivers/gpu/drm/panfrost/ |
D | panfrost_job.c | 43 to_panfrost_job(struct drm_sched_job *sched_job) in to_panfrost_job() argument 45 return container_of(sched_job, struct panfrost_job, base); in to_panfrost_job() 348 static void panfrost_job_free(struct drm_sched_job *sched_job) in panfrost_job_free() argument 350 struct panfrost_job *job = to_panfrost_job(sched_job); in panfrost_job_free() 352 drm_sched_job_cleanup(sched_job); in panfrost_job_free() 357 static struct dma_fence *panfrost_job_run(struct drm_sched_job *sched_job) in panfrost_job_run() argument 359 struct panfrost_job *job = to_panfrost_job(sched_job); in panfrost_job_run() 710 *sched_job) in panfrost_job_timedout() 712 struct panfrost_job *job = to_panfrost_job(sched_job); in panfrost_job_timedout() 729 sched_job); in panfrost_job_timedout() [all …]
|