Home
last modified time | relevance | path

Searched refs:sched_job (Results 1 – 13 of 13) sorted by relevance

/Linux-v6.1/drivers/gpu/drm/v3d/
Dv3d_sched.c28 to_v3d_job(struct drm_sched_job *sched_job) in to_v3d_job() argument
30 return container_of(sched_job, struct v3d_job, base); in to_v3d_job()
34 to_bin_job(struct drm_sched_job *sched_job) in to_bin_job() argument
36 return container_of(sched_job, struct v3d_bin_job, base.base); in to_bin_job()
40 to_render_job(struct drm_sched_job *sched_job) in to_render_job() argument
42 return container_of(sched_job, struct v3d_render_job, base.base); in to_render_job()
46 to_tfu_job(struct drm_sched_job *sched_job) in to_tfu_job() argument
48 return container_of(sched_job, struct v3d_tfu_job, base.base); in to_tfu_job()
52 to_csd_job(struct drm_sched_job *sched_job) in to_csd_job() argument
54 return container_of(sched_job, struct v3d_csd_job, base.base); in to_csd_job()
[all …]
/Linux-v6.1/drivers/gpu/drm/scheduler/
Dgpu_scheduler_trace.h36 TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
37 TP_ARGS(sched_job, entity),
41 __string(name, sched_job->sched->name)
49 __entry->id = sched_job->id;
50 __entry->fence = &sched_job->s_fence->finished;
51 __assign_str(name, sched_job->sched->name);
54 &sched_job->sched->hw_rq_count);
63 TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
64 TP_ARGS(sched_job, entity)
68 TP_PROTO(struct drm_sched_job *sched_job, struct drm_sched_entity *entity),
[all …]
Dsched_entity.c33 #define to_drm_sched_job(sched_job) \ argument
34 container_of((sched_job), struct drm_sched_job, queue_node)
421 struct drm_sched_job *sched_job; in drm_sched_entity_pop_job() local
423 sched_job = to_drm_sched_job(spsc_queue_peek(&entity->job_queue)); in drm_sched_entity_pop_job()
424 if (!sched_job) in drm_sched_entity_pop_job()
428 drm_sched_job_dependency(sched_job, entity))) { in drm_sched_entity_pop_job()
429 trace_drm_sched_job_wait_dep(sched_job, entity->dependency); in drm_sched_entity_pop_job()
437 dma_fence_set_error(&sched_job->s_fence->finished, -ECANCELED); in drm_sched_entity_pop_job()
441 entity->last_scheduled = dma_fence_get(&sched_job->s_fence->finished); in drm_sched_entity_pop_job()
451 return sched_job; in drm_sched_entity_pop_job()
[all …]
Dsched_main.c62 #define to_drm_sched_job(sched_job) \ argument
63 container_of((sched_job), struct drm_sched_job, queue_node)
927 struct drm_sched_job *sched_job; in drm_sched_main() local
943 sched_job = drm_sched_entity_pop_job(entity); in drm_sched_main()
945 if (!sched_job) { in drm_sched_main()
950 s_fence = sched_job->s_fence; in drm_sched_main()
953 drm_sched_job_begin(sched_job); in drm_sched_main()
955 trace_drm_run_job(sched_job, entity); in drm_sched_main()
956 fence = sched->ops->run_job(sched_job); in drm_sched_main()
965 r = dma_fence_add_callback(fence, &sched_job->cb, in drm_sched_main()
[all …]
/Linux-v6.1/drivers/gpu/drm/etnaviv/
Detnaviv_sched.c20 static struct dma_fence *etnaviv_sched_run_job(struct drm_sched_job *sched_job) in etnaviv_sched_run_job() argument
22 struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); in etnaviv_sched_run_job()
25 if (likely(!sched_job->s_fence->finished.error)) in etnaviv_sched_run_job()
34 *sched_job) in etnaviv_sched_timedout_job()
36 struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); in etnaviv_sched_timedout_job()
42 drm_sched_stop(&gpu->sched, sched_job); in etnaviv_sched_timedout_job()
65 if(sched_job) in etnaviv_sched_timedout_job()
66 drm_sched_increase_karma(sched_job); in etnaviv_sched_timedout_job()
83 static void etnaviv_sched_free_job(struct drm_sched_job *sched_job) in etnaviv_sched_free_job() argument
85 struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); in etnaviv_sched_free_job()
[all …]
Detnaviv_sched.h14 struct etnaviv_gem_submit *to_etnaviv_submit(struct drm_sched_job *sched_job) in to_etnaviv_submit() argument
16 return container_of(sched_job, struct etnaviv_gem_submit, sched_job); in to_etnaviv_submit()
Detnaviv_gem_submit.c189 ret = drm_sched_job_add_implicit_dependencies(&submit->sched_job, in submit_fence_sync()
532 ret = drm_sched_job_init(&submit->sched_job, in etnaviv_ioctl_gem_submit()
556 ret = drm_sched_job_add_dependency(&submit->sched_job, in etnaviv_ioctl_gem_submit()
616 drm_sched_job_cleanup(&submit->sched_job); in etnaviv_ioctl_gem_submit()
Detnaviv_gem.h90 struct drm_sched_job sched_job; member
/Linux-v6.1/include/drm/
Dgpu_scheduler.h358 struct dma_fence *(*dependency)(struct drm_sched_job *sched_job,
367 struct dma_fence *(*run_job)(struct drm_sched_job *sched_job);
411 enum drm_gpu_sched_stat (*timedout_job)(struct drm_sched_job *sched_job);
417 void (*free_job)(struct drm_sched_job *sched_job);
525 void drm_sched_entity_push_job(struct drm_sched_job *sched_job);
/Linux-v6.1/drivers/gpu/drm/amd/amdgpu/
Damdgpu_job.c245 static struct dma_fence *amdgpu_job_dependency(struct drm_sched_job *sched_job, in amdgpu_job_dependency() argument
249 struct amdgpu_job *job = to_amdgpu_job(sched_job); in amdgpu_job_dependency()
277 static struct dma_fence *amdgpu_job_run(struct drm_sched_job *sched_job) in amdgpu_job_run() argument
279 struct amdgpu_ring *ring = to_amdgpu_ring(sched_job->sched); in amdgpu_job_run()
285 job = to_amdgpu_job(sched_job); in amdgpu_job_run()
313 #define to_drm_sched_job(sched_job) \ argument
314 container_of((sched_job), struct drm_sched_job, queue_node)
Damdgpu_job.h39 #define to_amdgpu_job(sched_job) \ argument
40 container_of((sched_job), struct amdgpu_job, base)
Damdgpu_trace.h518 TP_PROTO(struct amdgpu_job *sched_job, struct dma_fence *fence),
519 TP_ARGS(sched_job, fence),
521 __string(ring, sched_job->base.sched->name)
529 __assign_str(ring, sched_job->base.sched->name);
530 __entry->id = sched_job->base.id;
/Linux-v6.1/drivers/gpu/drm/panfrost/
Dpanfrost_job.c43 to_panfrost_job(struct drm_sched_job *sched_job) in to_panfrost_job() argument
45 return container_of(sched_job, struct panfrost_job, base); in to_panfrost_job()
348 static void panfrost_job_free(struct drm_sched_job *sched_job) in panfrost_job_free() argument
350 struct panfrost_job *job = to_panfrost_job(sched_job); in panfrost_job_free()
352 drm_sched_job_cleanup(sched_job); in panfrost_job_free()
357 static struct dma_fence *panfrost_job_run(struct drm_sched_job *sched_job) in panfrost_job_run() argument
359 struct panfrost_job *job = to_panfrost_job(sched_job); in panfrost_job_run()
710 *sched_job) in panfrost_job_timedout()
712 struct panfrost_job *job = to_panfrost_job(sched_job); in panfrost_job_timedout()
729 sched_job); in panfrost_job_timedout()
[all …]