/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_job.c | 34 struct amdgpu_job *job = to_amdgpu_job(s_job); in amdgpu_job_timedout() local 37 job->base.sched->name, atomic_read(&ring->fence_drv.last_seq), in amdgpu_job_timedout() 40 amdgpu_device_gpu_recover(ring->adev, job, false); in amdgpu_job_timedout() 44 struct amdgpu_job **job, struct amdgpu_vm *vm) in amdgpu_job_alloc() argument 53 *job = kzalloc(size, GFP_KERNEL); in amdgpu_job_alloc() 54 if (!*job) in amdgpu_job_alloc() 61 (*job)->base.sched = &adev->rings[0]->sched; in amdgpu_job_alloc() 62 (*job)->vm = vm; in amdgpu_job_alloc() 63 (*job)->ibs = (void *)&(*job)[1]; in amdgpu_job_alloc() 64 (*job)->num_ibs = num_ibs; in amdgpu_job_alloc() [all …]
|
D | amdgpu_ib.c | 119 struct amdgpu_ib *ibs, struct amdgpu_job *job, in amdgpu_ib_schedule() argument 140 if (job) { in amdgpu_ib_schedule() 141 vm = job->vm; in amdgpu_ib_schedule() 142 fence_ctx = job->base.s_fence->scheduled.context; in amdgpu_ib_schedule() 153 if (vm && !job->vmid) { in amdgpu_ib_schedule() 168 if (ring->funcs->emit_pipeline_sync && job && in amdgpu_ib_schedule() 169 ((tmp = amdgpu_sync_get_fence(&job->sched_sync, NULL)) || in amdgpu_ib_schedule() 171 amdgpu_vm_need_pipeline_sync(ring, job))) { in amdgpu_ib_schedule() 179 if (job) { in amdgpu_ib_schedule() 180 r = amdgpu_vm_flush(ring, job, need_pipe_sync); in amdgpu_ib_schedule() [all …]
|
D | amdgpu_trace.h | 37 #define AMDGPU_JOB_GET_TIMELINE_NAME(job) \ argument 38 job->base.s_fence->finished.ops->get_timeline_name(&job->base.s_fence->finished) 154 __entry->dw = p->job->ibs[i].length_dw; 164 TP_PROTO(struct amdgpu_job *job), 165 TP_ARGS(job), 168 __string(timeline, AMDGPU_JOB_GET_TIMELINE_NAME(job)) 177 __entry->sched_job_id = job->base.id; 178 __assign_str(timeline, AMDGPU_JOB_GET_TIMELINE_NAME(job)) 179 __entry->context = job->base.s_fence->finished.context; 180 __entry->seqno = job->base.s_fence->finished.seqno; [all …]
|
D | amdgpu_job.h | 64 struct amdgpu_job **job, struct amdgpu_vm *vm); 66 struct amdgpu_job **job); 68 void amdgpu_job_free_resources(struct amdgpu_job *job); 69 void amdgpu_job_free(struct amdgpu_job *job); 70 int amdgpu_job_submit(struct amdgpu_job *job, struct drm_sched_entity *entity, 72 int amdgpu_job_submit_direct(struct amdgpu_job *job, struct amdgpu_ring *ring,
|
D | amdgpu_ids.c | 269 struct amdgpu_job *job, in amdgpu_vmid_grab_reserved() argument 286 job->vm_pd_addr != (*id)->pd_gpu_addr || in amdgpu_vmid_grab_reserved() 314 job->vm_needs_flush = needs_flush; in amdgpu_vmid_grab_reserved() 334 struct amdgpu_job *job, in amdgpu_vmid_grab_used() argument 344 job->vm_needs_flush = vm->use_cpu_for_update; in amdgpu_vmid_grab_used() 355 if ((*id)->pd_gpu_addr != job->vm_pd_addr) in amdgpu_vmid_grab_used() 383 job->vm_needs_flush |= needs_flush; in amdgpu_vmid_grab_used() 404 struct amdgpu_job *job) in amdgpu_vmid_grab() argument 419 r = amdgpu_vmid_grab_reserved(vm, ring, sync, fence, job, &id); in amdgpu_vmid_grab() 423 r = amdgpu_vmid_grab_used(vm, ring, sync, fence, job, &id); in amdgpu_vmid_grab() [all …]
|
/Linux-v4.19/drivers/md/ |
D | dm-kcopyd.c | 394 struct kcopyd_job *job; in pop_io_job() local 400 list_for_each_entry(job, jobs, list) { in pop_io_job() 401 if (job->rw == READ || !test_bit(DM_KCOPYD_WRITE_SEQ, &job->flags)) { in pop_io_job() 402 list_del(&job->list); in pop_io_job() 403 return job; in pop_io_job() 406 if (job->write_offset == job->master_job->write_offset) { in pop_io_job() 407 job->master_job->write_offset += job->source.count; in pop_io_job() 408 list_del(&job->list); in pop_io_job() 409 return job; in pop_io_job() 419 struct kcopyd_job *job = NULL; in pop() local [all …]
|
/Linux-v4.19/drivers/gpu/host1x/ |
D | job.c | 39 struct host1x_job *job = NULL; in host1x_job_alloc() local 54 mem = job = kzalloc(total, GFP_KERNEL); in host1x_job_alloc() 55 if (!job) in host1x_job_alloc() 58 kref_init(&job->ref); in host1x_job_alloc() 59 job->channel = ch; in host1x_job_alloc() 63 job->relocs = num_relocs ? mem : NULL; in host1x_job_alloc() 65 job->unpins = num_unpins ? mem : NULL; in host1x_job_alloc() 67 job->gathers = num_cmdbufs ? mem : NULL; in host1x_job_alloc() 69 job->addr_phys = num_unpins ? mem : NULL; in host1x_job_alloc() 71 job->reloc_addr_phys = job->addr_phys; in host1x_job_alloc() [all …]
|
D | cdma.c | 225 struct host1x_job *job) in cdma_start_timer_locked() argument 234 cdma->timeout.client = job->client; in cdma_start_timer_locked() 235 cdma->timeout.syncpt = host1x_syncpt_get(host, job->syncpt_id); in cdma_start_timer_locked() 236 cdma->timeout.syncpt_val = job->syncpt_end; in cdma_start_timer_locked() 240 msecs_to_jiffies(job->timeout)); in cdma_start_timer_locked() 267 struct host1x_job *job, *n; in update_cdma_locked() local 277 list_for_each_entry_safe(job, n, &cdma->sync_queue, list) { in update_cdma_locked() 279 host1x_syncpt_get(host1x, job->syncpt_id); in update_cdma_locked() 282 if (!host1x_syncpt_is_expired(sp, job->syncpt_end)) { in update_cdma_locked() 284 if (job->timeout) in update_cdma_locked() [all …]
|
/Linux-v4.19/block/ |
D | bsg-lib.c | 47 struct bsg_job *job = blk_mq_rq_to_pdu(rq); in bsg_transport_fill_hdr() local 49 job->request_len = hdr->request_len; in bsg_transport_fill_hdr() 50 job->request = memdup_user(uptr64(hdr->request), hdr->request_len); in bsg_transport_fill_hdr() 52 return PTR_ERR_OR_ZERO(job->request); in bsg_transport_fill_hdr() 57 struct bsg_job *job = blk_mq_rq_to_pdu(rq); in bsg_transport_complete_rq() local 64 hdr->device_status = job->result & 0xff; in bsg_transport_complete_rq() 65 hdr->transport_status = host_byte(job->result); in bsg_transport_complete_rq() 66 hdr->driver_status = driver_byte(job->result); in bsg_transport_complete_rq() 72 if (job->result < 0) { in bsg_transport_complete_rq() 74 job->reply_len = sizeof(u32); in bsg_transport_complete_rq() [all …]
|
/Linux-v4.19/drivers/gpu/drm/v3d/ |
D | v3d_sched.c | 36 struct v3d_job *job = to_v3d_job(sched_job); in v3d_job_free() local 38 v3d_exec_put(job->exec); in v3d_job_free() 49 struct v3d_job *job = to_v3d_job(sched_job); in v3d_job_dependency() local 50 struct v3d_exec_info *exec = job->exec; in v3d_job_dependency() 51 enum v3d_queue q = job == &exec->bin ? V3D_BIN : V3D_RENDER; in v3d_job_dependency() 54 fence = job->in_fence; in v3d_job_dependency() 56 job->in_fence = NULL; in v3d_job_dependency() 81 struct v3d_job *job = to_v3d_job(sched_job); in v3d_job_run() local 82 struct v3d_exec_info *exec = job->exec; in v3d_job_run() 83 enum v3d_queue q = job == &exec->bin ? V3D_BIN : V3D_RENDER; in v3d_job_run() [all …]
|
/Linux-v4.19/drivers/gpu/host1x/hw/ |
D | channel_hw.c | 61 static void submit_gathers(struct host1x_job *job) in submit_gathers() argument 63 struct host1x_cdma *cdma = &job->channel->cdma; in submit_gathers() 66 for (i = 0; i < job->num_gathers; i++) { in submit_gathers() 67 struct host1x_job_gather *g = &job->gathers[i]; in submit_gathers() 76 static inline void synchronize_syncpt_base(struct host1x_job *job) in synchronize_syncpt_base() argument 78 struct host1x *host = dev_get_drvdata(job->channel->dev->parent); in synchronize_syncpt_base() 79 struct host1x_syncpt *sp = host->syncpt + job->syncpt_id; in synchronize_syncpt_base() 86 host1x_cdma_push(&job->channel->cdma, in synchronize_syncpt_base() 93 static int channel_submit(struct host1x_job *job) in channel_submit() argument 95 struct host1x_channel *ch = job->channel; in channel_submit() [all …]
|
D | debug_hw.c | 204 struct host1x_job *job; in show_channel_gathers() local 206 list_for_each_entry(job, &cdma->sync_queue, list) { in show_channel_gathers() 210 job, job->syncpt_id, job->syncpt_end, in show_channel_gathers() 211 job->first_get, job->timeout, in show_channel_gathers() 212 job->num_slots, job->num_unpins); in show_channel_gathers() 214 for (i = 0; i < job->num_gathers; i++) { in show_channel_gathers() 215 struct host1x_job_gather *g = &job->gathers[i]; in show_channel_gathers() 218 if (job->gather_copy_mapped) in show_channel_gathers() 219 mapped = (u32 *)job->gather_copy_mapped; in show_channel_gathers() 234 if (!job->gather_copy_mapped) in show_channel_gathers()
|
/Linux-v4.19/drivers/gpu/drm/ |
D | drm_writeback.c | 261 struct drm_writeback_job *job) in drm_writeback_queue_job() argument 266 list_add_tail(&job->list_entry, &wb_connector->job_queue); in drm_writeback_queue_job() 280 struct drm_writeback_job *job = container_of(work, in cleanup_work() local 283 drm_framebuffer_put(job->fb); in cleanup_work() 284 kfree(job); in cleanup_work() 309 struct drm_writeback_job *job; in drm_writeback_signal_completion() local 312 job = list_first_entry_or_null(&wb_connector->job_queue, in drm_writeback_signal_completion() 315 if (job) { in drm_writeback_signal_completion() 316 list_del(&job->list_entry); in drm_writeback_signal_completion() 317 if (job->out_fence) { in drm_writeback_signal_completion() [all …]
|
/Linux-v4.19/drivers/scsi/lpfc/ |
D | lpfc_bsg.c | 303 struct bsg_job *job; in lpfc_bsg_send_mgmt_cmd_cmp() local 317 job = dd_data->set_job; in lpfc_bsg_send_mgmt_cmd_cmp() 318 if (job) { in lpfc_bsg_send_mgmt_cmd_cmp() 319 bsg_reply = job->reply; in lpfc_bsg_send_mgmt_cmd_cmp() 321 job->dd_data = NULL; in lpfc_bsg_send_mgmt_cmd_cmp() 339 if (job) { in lpfc_bsg_send_mgmt_cmd_cmp() 359 lpfc_bsg_copy_data(rmp, &job->reply_payload, in lpfc_bsg_send_mgmt_cmd_cmp() 374 if (job) { in lpfc_bsg_send_mgmt_cmd_cmp() 376 bsg_job_done(job, bsg_reply->result, in lpfc_bsg_send_mgmt_cmd_cmp() 387 lpfc_bsg_send_mgmt_cmd(struct bsg_job *job) in lpfc_bsg_send_mgmt_cmd() argument [all …]
|
/Linux-v4.19/Documentation/devicetree/bindings/powerpc/fsl/ |
D | raideng.txt | 30 There must be a sub-node for each job queue present in RAID Engine 33 - compatible: Should contain "fsl,raideng-v1.0-job-queue" as the value 34 This identifies the job queue interface 35 - reg: offset and length of the register set for job queue 42 compatible = "fsl,raideng-v1.0-job-queue"; 48 There must be a sub-node for each job ring present in RAID Engine 49 This node must be a sub-node of job queue node 51 - compatible: Must contain "fsl,raideng-v1.0-job-ring" as the value 52 This identifies job ring. Should contain either 55 - reg: offset and length of the register set for job ring [all …]
|
/Linux-v4.19/drivers/scsi/libsas/ |
D | sas_host_smp.c | 228 void sas_smp_host_handler(struct bsg_job *job, struct Scsi_Host *shost) in sas_smp_host_handler() argument 236 if (job->request_payload.payload_len < 8 || in sas_smp_host_handler() 237 job->reply_payload.payload_len < 8) in sas_smp_host_handler() 241 req_data = kzalloc(job->request_payload.payload_len, GFP_KERNEL); in sas_smp_host_handler() 244 sg_copy_to_buffer(job->request_payload.sg_list, in sas_smp_host_handler() 245 job->request_payload.sg_cnt, req_data, in sas_smp_host_handler() 246 job->request_payload.payload_len); in sas_smp_host_handler() 250 resp_data = kzalloc(max(job->reply_payload.payload_len, 128U), in sas_smp_host_handler() 285 if (job->request_payload.payload_len < 16) in sas_smp_host_handler() 297 if (job->request_payload.payload_len < 16) in sas_smp_host_handler() [all …]
|
/Linux-v4.19/arch/x86/crypto/sha512-mb/ |
D | sha512_mb_mgr_submit_avx2.S | 71 #define job arg2 macro 98 # JOB* sha512_mb_mgr_submit_avx2(MB_MGR *state, JOB *job) 100 # arg 2 : rdx : job 110 movl $STS_BEING_PROCESSED, _status(job) 113 movl _len(job), DWORD_len 115 mov job, _job_in_lane(lane_data) 119 vmovdqu _result_digest+0*16(job), %xmm0 120 vmovdqu _result_digest+1*16(job), %xmm1 121 vmovdqu _result_digest+2*16(job), %xmm2 122 vmovdqu _result_digest+3*16(job), %xmm3 [all …]
|
/Linux-v4.19/drivers/gpu/drm/scheduler/ |
D | gpu_scheduler.c | 240 struct drm_sched_job *job = container_of(cb, struct drm_sched_job, in drm_sched_entity_kill_jobs_cb() local 242 drm_sched_fence_finished(job->s_fence); in drm_sched_entity_kill_jobs_cb() 243 WARN_ON(job->s_fence->parent); in drm_sched_entity_kill_jobs_cb() 244 dma_fence_put(&job->s_fence->finished); in drm_sched_entity_kill_jobs_cb() 245 job->sched->ops->free_job(job); in drm_sched_entity_kill_jobs_cb() 311 struct drm_sched_job *job; in drm_sched_entity_fini() local 326 while ((job = to_drm_sched_job(spsc_queue_pop(&entity->job_queue)))) { in drm_sched_entity_fini() 327 struct drm_sched_fence *s_fence = job->s_fence; in drm_sched_entity_fini() 337 drm_sched_entity_kill_jobs_cb(NULL, &job->finish_cb); in drm_sched_entity_fini() 339 r = dma_fence_add_callback(entity->last_scheduled, &job->finish_cb, in drm_sched_entity_fini() [all …]
|
/Linux-v4.19/arch/powerpc/boot/dts/fsl/ |
D | qoriq-sec6.0-0.dtsi | 42 compatible = "fsl,sec-v6.0-job-ring", 43 "fsl,sec-v5.2-job-ring", 44 "fsl,sec-v5.0-job-ring", 45 "fsl,sec-v4.4-job-ring", 46 "fsl,sec-v4.0-job-ring"; 51 compatible = "fsl,sec-v6.0-job-ring", 52 "fsl,sec-v5.2-job-ring", 53 "fsl,sec-v5.0-job-ring", 54 "fsl,sec-v4.4-job-ring", 55 "fsl,sec-v4.0-job-ring";
|
D | qoriq-sec5.3-0.dtsi | 45 compatible = "fsl,sec-v5.3-job-ring", 46 "fsl,sec-v5.0-job-ring", 47 "fsl,sec-v4.0-job-ring"; 53 compatible = "fsl,sec-v5.3-job-ring", 54 "fsl,sec-v5.0-job-ring", 55 "fsl,sec-v4.0-job-ring"; 61 compatible = "fsl,sec-v5.3-job-ring", 62 "fsl,sec-v5.0-job-ring", 63 "fsl,sec-v4.0-job-ring"; 69 compatible = "fsl,sec-v5.3-job-ring", [all …]
|
D | qoriq-sec5.2-0.dtsi | 45 compatible = "fsl,sec-v5.2-job-ring", 46 "fsl,sec-v5.0-job-ring", 47 "fsl,sec-v4.0-job-ring"; 53 compatible = "fsl,sec-v5.2-job-ring", 54 "fsl,sec-v5.0-job-ring", 55 "fsl,sec-v4.0-job-ring"; 61 compatible = "fsl,sec-v5.2-job-ring", 62 "fsl,sec-v5.0-job-ring", 63 "fsl,sec-v4.0-job-ring"; 69 compatible = "fsl,sec-v5.2-job-ring", [all …]
|
/Linux-v4.19/drivers/media/platform/ |
D | rcar_fdp1.c | 653 struct fdp1_job *job; in list_remove_job() local 657 job = list_first_entry_or_null(list, struct fdp1_job, list); in list_remove_job() 658 if (job) in list_remove_job() 659 list_del(&job->list); in list_remove_job() 662 return job; in list_remove_job() 672 struct fdp1_job *job) in list_add_job() argument 677 list_add_tail(&job->list, list); in list_add_job() 686 static void fdp1_job_free(struct fdp1_dev *fdp1, struct fdp1_job *job) in fdp1_job_free() argument 689 memset(job, 0, sizeof(struct fdp1_job)); in fdp1_job_free() 691 list_add_job(fdp1, &fdp1->free_job_list, job); in fdp1_job_free() [all …]
|
/Linux-v4.19/arch/x86/crypto/sha1-mb/ |
D | sha1_mb_mgr_submit_avx2.S | 71 #define job %rsi macro 98 # JOB* submit_mb_mgr_submit_avx2(MB_MGR *state, job_sha1 *job) 100 # arg 2 : rdx : job 111 movl $STS_BEING_PROCESSED, _status(job) 114 movl _len(job), DWORD_len 116 mov job, _job_in_lane(lane_data) 123 vmovdqu _result_digest(job), %xmm0 124 mov _result_digest+1*16(job), DWORD_tmp 131 mov _buffer(job), p 171 # process completed job "idx"
|
/Linux-v4.19/Documentation/devicetree/bindings/crypto/ |
D | fsl-sec6.txt | 84 Definition: Must include "fsl,sec-v6.0-job-ring". 103 compatible = "fsl,sec-v6.0-job-ring"; 123 compatible = "fsl,sec-v6.0-job-ring", 124 "fsl,sec-v5.2-job-ring", 125 "fsl,sec-v5.0-job-ring", 126 "fsl,sec-v4.4-job-ring", 127 "fsl,sec-v4.0-job-ring"; 132 compatible = "fsl,sec-v6.0-job-ring", 133 "fsl,sec-v5.2-job-ring", 134 "fsl,sec-v5.0-job-ring", [all …]
|
/Linux-v4.19/arch/x86/crypto/sha256-mb/ |
D | sha256_mb_mgr_submit_avx2.S | 69 #define job %rsi macro 96 # JOB* sha256_mb_mgr_submit_avx2(MB_MGR *state, JOB_SHA256 *job) 98 # arg 2 : rdx : job 109 movl $STS_BEING_PROCESSED, _status(job) 112 movl _len(job), DWORD_len 114 mov job, _job_in_lane(lane_data) 121 vmovdqu _result_digest(job), %xmm0 122 vmovdqu _result_digest+1*16(job), %xmm1 133 mov _buffer(job), p 172 # process completed job "idx"
|