Home
last modified time | relevance | path

Searched refs:job (Results 1 – 25 of 173) sorted by relevance

1234567

/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/
Damdgpu_job.c34 struct amdgpu_job *job = to_amdgpu_job(s_job); in amdgpu_job_timedout() local
37 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout()
40 amdgpu_device_gpu_recover(ring->adev, job, false); in amdgpu_job_timedout()
44 struct amdgpu_job **job, struct amdgpu_vm *vm) in amdgpu_job_alloc() argument
53 *job = kzalloc(size, GFP_KERNEL); in amdgpu_job_alloc()
54 if (!*job) in amdgpu_job_alloc()
61 (*job)->base.sched = &adev->rings[0]->sched; in amdgpu_job_alloc()
62 (*job)->vm = vm; in amdgpu_job_alloc()
63 (*job)->ibs = (void *)&(*job)[1]; in amdgpu_job_alloc()
64 (*job)->num_ibs = num_ibs; in amdgpu_job_alloc()
[all …]
Damdgpu_ib.c119 struct amdgpu_ib *ibs, struct amdgpu_job *job, in amdgpu_ib_schedule() argument
140 if (job) { in amdgpu_ib_schedule()
141 vm = job->vm; in amdgpu_ib_schedule()
142 fence_ctx = job->base.s_fence->scheduled.context; in amdgpu_ib_schedule()
153 if (vm && !job->vmid) { in amdgpu_ib_schedule()
168 if (ring->funcs->emit_pipeline_sync && job && in amdgpu_ib_schedule()
169 ((tmp = amdgpu_sync_get_fence(&job->sched_sync, NULL)) || in amdgpu_ib_schedule()
171 amdgpu_vm_need_pipeline_sync(ring, job))) { in amdgpu_ib_schedule()
179 if (job) { in amdgpu_ib_schedule()
180 r = amdgpu_vm_flush(ring, job, need_pipe_sync); in amdgpu_ib_schedule()
[all …]
Damdgpu_trace.h37 #define AMDGPU_JOB_GET_TIMELINE_NAME(job) \ argument
38 job->base.s_fence->finished.ops->get_timeline_name(&job->base.s_fence->finished)
154 __entry->dw = p->job->ibs[i].length_dw;
164 TP_PROTO(struct amdgpu_job *job),
165 TP_ARGS(job),
168 __string(timeline, AMDGPU_JOB_GET_TIMELINE_NAME(job))
177 __entry->sched_job_id = job->base.id;
178 __assign_str(timeline, AMDGPU_JOB_GET_TIMELINE_NAME(job))
179 __entry->context = job->base.s_fence->finished.context;
180 __entry->seqno = job->base.s_fence->finished.seqno;
[all …]
Damdgpu_job.h64 struct amdgpu_job **job, struct amdgpu_vm *vm);
66 struct amdgpu_job **job);
68 void amdgpu_job_free_resources(struct amdgpu_job *job);
69 void amdgpu_job_free(struct amdgpu_job *job);
70 int amdgpu_job_submit(struct amdgpu_job *job, struct drm_sched_entity *entity,
72 int amdgpu_job_submit_direct(struct amdgpu_job *job, struct amdgpu_ring *ring,
Damdgpu_ids.c269 struct amdgpu_job *job, in amdgpu_vmid_grab_reserved() argument
286 job->vm_pd_addr != (*id)->pd_gpu_addr || in amdgpu_vmid_grab_reserved()
314 job->vm_needs_flush = needs_flush; in amdgpu_vmid_grab_reserved()
334 struct amdgpu_job *job, in amdgpu_vmid_grab_used() argument
344 job->vm_needs_flush = vm->use_cpu_for_update; in amdgpu_vmid_grab_used()
355 if ((*id)->pd_gpu_addr != job->vm_pd_addr) in amdgpu_vmid_grab_used()
383 job->vm_needs_flush |= needs_flush; in amdgpu_vmid_grab_used()
404 struct amdgpu_job *job) in amdgpu_vmid_grab() argument
419 r = amdgpu_vmid_grab_reserved(vm, ring, sync, fence, job, &id); in amdgpu_vmid_grab()
423 r = amdgpu_vmid_grab_used(vm, ring, sync, fence, job, &id); in amdgpu_vmid_grab()
[all …]
/Linux-v4.19/drivers/md/
Ddm-kcopyd.c394 struct kcopyd_job *job; in pop_io_job() local
400 list_for_each_entry(job, jobs, list) { in pop_io_job()
401 if (job->rw == READ || !test_bit(DM_KCOPYD_WRITE_SEQ, &job->flags)) { in pop_io_job()
402 list_del(&job->list); in pop_io_job()
403 return job; in pop_io_job()
406 if (job->write_offset == job->master_job->write_offset) { in pop_io_job()
407 job->master_job->write_offset += job->source.count; in pop_io_job()
408 list_del(&job->list); in pop_io_job()
409 return job; in pop_io_job()
419 struct kcopyd_job *job = NULL; in pop() local
[all …]
/Linux-v4.19/drivers/gpu/host1x/
Djob.c39 struct host1x_job *job = NULL; in host1x_job_alloc() local
54 mem = job = kzalloc(total, GFP_KERNEL); in host1x_job_alloc()
55 if (!job) in host1x_job_alloc()
58 kref_init(&job->ref); in host1x_job_alloc()
59 job->channel = ch; in host1x_job_alloc()
63 job->relocs = num_relocs ? mem : NULL; in host1x_job_alloc()
65 job->unpins = num_unpins ? mem : NULL; in host1x_job_alloc()
67 job->gathers = num_cmdbufs ? mem : NULL; in host1x_job_alloc()
69 job->addr_phys = num_unpins ? mem : NULL; in host1x_job_alloc()
71 job->reloc_addr_phys = job->addr_phys; in host1x_job_alloc()
[all …]
Dcdma.c225 struct host1x_job *job) in cdma_start_timer_locked() argument
234 cdma->timeout.client = job->client; in cdma_start_timer_locked()
235 cdma->timeout.syncpt = host1x_syncpt_get(host, job->syncpt_id); in cdma_start_timer_locked()
236 cdma->timeout.syncpt_val = job->syncpt_end; in cdma_start_timer_locked()
240 msecs_to_jiffies(job->timeout)); in cdma_start_timer_locked()
267 struct host1x_job *job, *n; in update_cdma_locked() local
277 list_for_each_entry_safe(job, n, &cdma->sync_queue, list) { in update_cdma_locked()
279 host1x_syncpt_get(host1x, job->syncpt_id); in update_cdma_locked()
282 if (!host1x_syncpt_is_expired(sp, job->syncpt_end)) { in update_cdma_locked()
284 if (job->timeout) in update_cdma_locked()
[all …]
/Linux-v4.19/block/
Dbsg-lib.c47 struct bsg_job *job = blk_mq_rq_to_pdu(rq); in bsg_transport_fill_hdr() local
49 job->request_len = hdr->request_len; in bsg_transport_fill_hdr()
50 job->request = memdup_user(uptr64(hdr->request), hdr->request_len); in bsg_transport_fill_hdr()
52 return PTR_ERR_OR_ZERO(job->request); in bsg_transport_fill_hdr()
57 struct bsg_job *job = blk_mq_rq_to_pdu(rq); in bsg_transport_complete_rq() local
64 hdr->device_status = job->result & 0xff; in bsg_transport_complete_rq()
65 hdr->transport_status = host_byte(job->result); in bsg_transport_complete_rq()
66 hdr->driver_status = driver_byte(job->result); in bsg_transport_complete_rq()
72 if (job->result < 0) { in bsg_transport_complete_rq()
74 job->reply_len = sizeof(u32); in bsg_transport_complete_rq()
[all …]
/Linux-v4.19/drivers/gpu/drm/v3d/
Dv3d_sched.c36 struct v3d_job *job = to_v3d_job(sched_job); in v3d_job_free() local
38 v3d_exec_put(job->exec); in v3d_job_free()
49 struct v3d_job *job = to_v3d_job(sched_job); in v3d_job_dependency() local
50 struct v3d_exec_info *exec = job->exec; in v3d_job_dependency()
51 enum v3d_queue q = job == &exec->bin ? V3D_BIN : V3D_RENDER; in v3d_job_dependency()
54 fence = job->in_fence; in v3d_job_dependency()
56 job->in_fence = NULL; in v3d_job_dependency()
81 struct v3d_job *job = to_v3d_job(sched_job); in v3d_job_run() local
82 struct v3d_exec_info *exec = job->exec; in v3d_job_run()
83 enum v3d_queue q = job == &exec->bin ? V3D_BIN : V3D_RENDER; in v3d_job_run()
[all …]
/Linux-v4.19/drivers/gpu/host1x/hw/
Dchannel_hw.c61 static void submit_gathers(struct host1x_job *job) in submit_gathers() argument
63 struct host1x_cdma *cdma = &job->channel->cdma; in submit_gathers()
66 for (i = 0; i < job->num_gathers; i++) { in submit_gathers()
67 struct host1x_job_gather *g = &job->gathers[i]; in submit_gathers()
76 static inline void synchronize_syncpt_base(struct host1x_job *job) in synchronize_syncpt_base() argument
78 struct host1x *host = dev_get_drvdata(job->channel->dev->parent); in synchronize_syncpt_base()
79 struct host1x_syncpt *sp = host->syncpt + job->syncpt_id; in synchronize_syncpt_base()
86 host1x_cdma_push(&job->channel->cdma, in synchronize_syncpt_base()
93 static int channel_submit(struct host1x_job *job) in channel_submit() argument
95 struct host1x_channel *ch = job->channel; in channel_submit()
[all …]
Ddebug_hw.c204 struct host1x_job *job; in show_channel_gathers() local
206 list_for_each_entry(job, &cdma->sync_queue, list) { in show_channel_gathers()
210 job, job->syncpt_id, job->syncpt_end, in show_channel_gathers()
211 job->first_get, job->timeout, in show_channel_gathers()
212 job->num_slots, job->num_unpins); in show_channel_gathers()
214 for (i = 0; i < job->num_gathers; i++) { in show_channel_gathers()
215 struct host1x_job_gather *g = &job->gathers[i]; in show_channel_gathers()
218 if (job->gather_copy_mapped) in show_channel_gathers()
219 mapped = (u32 *)job->gather_copy_mapped; in show_channel_gathers()
234 if (!job->gather_copy_mapped) in show_channel_gathers()
/Linux-v4.19/drivers/gpu/drm/
Ddrm_writeback.c261 struct drm_writeback_job *job) in drm_writeback_queue_job() argument
266 list_add_tail(&job->list_entry, &wb_connector->job_queue); in drm_writeback_queue_job()
280 struct drm_writeback_job *job = container_of(work, in cleanup_work() local
283 drm_framebuffer_put(job->fb); in cleanup_work()
284 kfree(job); in cleanup_work()
309 struct drm_writeback_job *job; in drm_writeback_signal_completion() local
312 job = list_first_entry_or_null(&wb_connector->job_queue, in drm_writeback_signal_completion()
315 if (job) { in drm_writeback_signal_completion()
316 list_del(&job->list_entry); in drm_writeback_signal_completion()
317 if (job->out_fence) { in drm_writeback_signal_completion()
[all …]
/Linux-v4.19/drivers/scsi/lpfc/
Dlpfc_bsg.c303 struct bsg_job *job; in lpfc_bsg_send_mgmt_cmd_cmp() local
317 job = dd_data->set_job; in lpfc_bsg_send_mgmt_cmd_cmp()
318 if (job) { in lpfc_bsg_send_mgmt_cmd_cmp()
319 bsg_reply = job->reply; in lpfc_bsg_send_mgmt_cmd_cmp()
321 job->dd_data = NULL; in lpfc_bsg_send_mgmt_cmd_cmp()
339 if (job) { in lpfc_bsg_send_mgmt_cmd_cmp()
359 lpfc_bsg_copy_data(rmp, &job->reply_payload, in lpfc_bsg_send_mgmt_cmd_cmp()
374 if (job) { in lpfc_bsg_send_mgmt_cmd_cmp()
376 bsg_job_done(job, bsg_reply->result, in lpfc_bsg_send_mgmt_cmd_cmp()
387 lpfc_bsg_send_mgmt_cmd(struct bsg_job *job) in lpfc_bsg_send_mgmt_cmd() argument
[all …]
/Linux-v4.19/Documentation/devicetree/bindings/powerpc/fsl/
Draideng.txt30 There must be a sub-node for each job queue present in RAID Engine
33 - compatible: Should contain "fsl,raideng-v1.0-job-queue" as the value
34 This identifies the job queue interface
35 - reg: offset and length of the register set for job queue
42 compatible = "fsl,raideng-v1.0-job-queue";
48 There must be a sub-node for each job ring present in RAID Engine
49 This node must be a sub-node of job queue node
51 - compatible: Must contain "fsl,raideng-v1.0-job-ring" as the value
52 This identifies job ring. Should contain either
55 - reg: offset and length of the register set for job ring
[all …]
/Linux-v4.19/drivers/scsi/libsas/
Dsas_host_smp.c228 void sas_smp_host_handler(struct bsg_job *job, struct Scsi_Host *shost) in sas_smp_host_handler() argument
236 if (job->request_payload.payload_len < 8 || in sas_smp_host_handler()
237 job->reply_payload.payload_len < 8) in sas_smp_host_handler()
241 req_data = kzalloc(job->request_payload.payload_len, GFP_KERNEL); in sas_smp_host_handler()
244 sg_copy_to_buffer(job->request_payload.sg_list, in sas_smp_host_handler()
245 job->request_payload.sg_cnt, req_data, in sas_smp_host_handler()
246 job->request_payload.payload_len); in sas_smp_host_handler()
250 resp_data = kzalloc(max(job->reply_payload.payload_len, 128U), in sas_smp_host_handler()
285 if (job->request_payload.payload_len < 16) in sas_smp_host_handler()
297 if (job->request_payload.payload_len < 16) in sas_smp_host_handler()
[all …]
/Linux-v4.19/arch/x86/crypto/sha512-mb/
Dsha512_mb_mgr_submit_avx2.S71 #define job arg2 macro
98 # JOB* sha512_mb_mgr_submit_avx2(MB_MGR *state, JOB *job)
100 # arg 2 : rdx : job
110 movl $STS_BEING_PROCESSED, _status(job)
113 movl _len(job), DWORD_len
115 mov job, _job_in_lane(lane_data)
119 vmovdqu _result_digest+0*16(job), %xmm0
120 vmovdqu _result_digest+1*16(job), %xmm1
121 vmovdqu _result_digest+2*16(job), %xmm2
122 vmovdqu _result_digest+3*16(job), %xmm3
[all …]
/Linux-v4.19/drivers/gpu/drm/scheduler/
Dgpu_scheduler.c240 struct drm_sched_job *job = container_of(cb, struct drm_sched_job, in drm_sched_entity_kill_jobs_cb() local
242 drm_sched_fence_finished(job->s_fence); in drm_sched_entity_kill_jobs_cb()
243 WARN_ON(job->s_fence->parent); in drm_sched_entity_kill_jobs_cb()
244 dma_fence_put(&job->s_fence->finished); in drm_sched_entity_kill_jobs_cb()
245 job->sched->ops->free_job(job); in drm_sched_entity_kill_jobs_cb()
311 struct drm_sched_job *job; in drm_sched_entity_fini() local
326 while ((job = to_drm_sched_job(spsc_queue_pop(&entity->job_queue)))) { in drm_sched_entity_fini()
327 struct drm_sched_fence *s_fence = job->s_fence; in drm_sched_entity_fini()
337 drm_sched_entity_kill_jobs_cb(NULL, &job->finish_cb); in drm_sched_entity_fini()
339 r = dma_fence_add_callback(entity->last_scheduled, &job->finish_cb, in drm_sched_entity_fini()
[all …]
/Linux-v4.19/arch/powerpc/boot/dts/fsl/
Dqoriq-sec6.0-0.dtsi42 compatible = "fsl,sec-v6.0-job-ring",
43 "fsl,sec-v5.2-job-ring",
44 "fsl,sec-v5.0-job-ring",
45 "fsl,sec-v4.4-job-ring",
46 "fsl,sec-v4.0-job-ring";
51 compatible = "fsl,sec-v6.0-job-ring",
52 "fsl,sec-v5.2-job-ring",
53 "fsl,sec-v5.0-job-ring",
54 "fsl,sec-v4.4-job-ring",
55 "fsl,sec-v4.0-job-ring";
Dqoriq-sec5.3-0.dtsi45 compatible = "fsl,sec-v5.3-job-ring",
46 "fsl,sec-v5.0-job-ring",
47 "fsl,sec-v4.0-job-ring";
53 compatible = "fsl,sec-v5.3-job-ring",
54 "fsl,sec-v5.0-job-ring",
55 "fsl,sec-v4.0-job-ring";
61 compatible = "fsl,sec-v5.3-job-ring",
62 "fsl,sec-v5.0-job-ring",
63 "fsl,sec-v4.0-job-ring";
69 compatible = "fsl,sec-v5.3-job-ring",
[all …]
Dqoriq-sec5.2-0.dtsi45 compatible = "fsl,sec-v5.2-job-ring",
46 "fsl,sec-v5.0-job-ring",
47 "fsl,sec-v4.0-job-ring";
53 compatible = "fsl,sec-v5.2-job-ring",
54 "fsl,sec-v5.0-job-ring",
55 "fsl,sec-v4.0-job-ring";
61 compatible = "fsl,sec-v5.2-job-ring",
62 "fsl,sec-v5.0-job-ring",
63 "fsl,sec-v4.0-job-ring";
69 compatible = "fsl,sec-v5.2-job-ring",
[all …]
/Linux-v4.19/drivers/media/platform/
Drcar_fdp1.c653 struct fdp1_job *job; in list_remove_job() local
657 job = list_first_entry_or_null(list, struct fdp1_job, list); in list_remove_job()
658 if (job) in list_remove_job()
659 list_del(&job->list); in list_remove_job()
662 return job; in list_remove_job()
672 struct fdp1_job *job) in list_add_job() argument
677 list_add_tail(&job->list, list); in list_add_job()
686 static void fdp1_job_free(struct fdp1_dev *fdp1, struct fdp1_job *job) in fdp1_job_free() argument
689 memset(job, 0, sizeof(struct fdp1_job)); in fdp1_job_free()
691 list_add_job(fdp1, &fdp1->free_job_list, job); in fdp1_job_free()
[all …]
/Linux-v4.19/arch/x86/crypto/sha1-mb/
Dsha1_mb_mgr_submit_avx2.S71 #define job %rsi macro
98 # JOB* submit_mb_mgr_submit_avx2(MB_MGR *state, job_sha1 *job)
100 # arg 2 : rdx : job
111 movl $STS_BEING_PROCESSED, _status(job)
114 movl _len(job), DWORD_len
116 mov job, _job_in_lane(lane_data)
123 vmovdqu _result_digest(job), %xmm0
124 mov _result_digest+1*16(job), DWORD_tmp
131 mov _buffer(job), p
171 # process completed job "idx"
/Linux-v4.19/Documentation/devicetree/bindings/crypto/
Dfsl-sec6.txt84 Definition: Must include "fsl,sec-v6.0-job-ring".
103 compatible = "fsl,sec-v6.0-job-ring";
123 compatible = "fsl,sec-v6.0-job-ring",
124 "fsl,sec-v5.2-job-ring",
125 "fsl,sec-v5.0-job-ring",
126 "fsl,sec-v4.4-job-ring",
127 "fsl,sec-v4.0-job-ring";
132 compatible = "fsl,sec-v6.0-job-ring",
133 "fsl,sec-v5.2-job-ring",
134 "fsl,sec-v5.0-job-ring",
[all …]
/Linux-v4.19/arch/x86/crypto/sha256-mb/
Dsha256_mb_mgr_submit_avx2.S69 #define job %rsi macro
96 # JOB* sha256_mb_mgr_submit_avx2(MB_MGR *state, JOB_SHA256 *job)
98 # arg 2 : rdx : job
109 movl $STS_BEING_PROCESSED, _status(job)
112 movl _len(job), DWORD_len
114 mov job, _job_in_lane(lane_data)
121 vmovdqu _result_digest(job), %xmm0
122 vmovdqu _result_digest+1*16(job), %xmm1
133 mov _buffer(job), p
172 # process completed job "idx"

1234567