/Linux-v4.19/drivers/gpu/drm/msm/ |
D | msm_gem_submit.c | 37 struct msm_gem_submit *submit; in submit_create() local 38 uint64_t sz = sizeof(*submit) + ((u64)nr_bos * sizeof(submit->bos[0])) + in submit_create() 39 ((u64)nr_cmds * sizeof(submit->cmd[0])); in submit_create() 44 submit = kmalloc(sz, GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY); in submit_create() 45 if (!submit) in submit_create() 48 submit->dev = dev; in submit_create() 49 submit->gpu = gpu; in submit_create() 50 submit->fence = NULL; in submit_create() 51 submit->pid = get_pid(task_pid(current)); in submit_create() 52 submit->cmd = (void *)&submit->bos[nr_bos]; in submit_create() [all …]
|
D | msm_rd.c | 94 struct msm_gem_submit *submit; member 315 struct msm_gem_submit *submit, int idx, in snapshot_buf() argument 318 struct msm_gem_object *obj = submit->bos[idx].obj; in snapshot_buf() 322 buf += iova - submit->bos[idx].iova; in snapshot_buf() 324 iova = submit->bos[idx].iova; in snapshot_buf() 336 if (!(submit->bos[idx].flags & MSM_SUBMIT_BO_READ)) in snapshot_buf() 349 void msm_rd_dump_submit(struct msm_rd_state *rd, struct msm_gem_submit *submit, in msm_rd_dump_submit() argument 352 struct drm_device *dev = submit->dev; in msm_rd_dump_submit() 376 task = pid_task(submit->pid, PIDTYPE_PID); in msm_rd_dump_submit() 380 pid_nr(submit->pid), submit->seqno); in msm_rd_dump_submit() [all …]
|
D | msm_gpu.c | 335 struct msm_gem_submit *submit, char *comm, char *cmd) in msm_gpu_crashstate_capture() argument 351 if (submit) { in msm_gpu_crashstate_capture() 354 state->bos = kcalloc(submit->nr_bos, in msm_gpu_crashstate_capture() 357 for (i = 0; state->bos && i < submit->nr_bos; i++) in msm_gpu_crashstate_capture() 358 msm_gpu_crashstate_get_bo(state, submit->bos[i].obj, in msm_gpu_crashstate_capture() 359 submit->bos[i].iova, submit->bos[i].flags); in msm_gpu_crashstate_capture() 383 struct msm_gem_submit *submit; in update_fences() local 385 list_for_each_entry(submit, &ring->submits, node) { in update_fences() 386 if (submit->seqno > fence) in update_fences() 389 msm_update_fence(submit->ring->fctx, in update_fences() [all …]
|
D | msm_gpu.h | 59 void (*submit)(struct msm_gpu *gpu, struct msm_gem_submit *submit, member 276 void msm_gpu_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit,
|
/Linux-v4.19/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem_submit.c | 39 struct etnaviv_gem_submit *submit; in submit_create() local 40 size_t sz = size_vstruct(nr_bos, sizeof(submit->bos[0]), sizeof(*submit)); in submit_create() 42 submit = kzalloc(sz, GFP_KERNEL); in submit_create() 43 if (!submit) in submit_create() 46 submit->pmrs = kcalloc(nr_pmrs, sizeof(struct etnaviv_perfmon_request), in submit_create() 48 if (!submit->pmrs) { in submit_create() 49 kfree(submit); in submit_create() 52 submit->nr_pmrs = nr_pmrs; in submit_create() 54 submit->gpu = gpu; in submit_create() 55 kref_init(&submit->refcount); in submit_create() [all …]
|
D | etnaviv_sched.c | 24 struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); in etnaviv_sched_dependency() local 28 if (unlikely(submit->in_fence)) { in etnaviv_sched_dependency() 29 fence = submit->in_fence; in etnaviv_sched_dependency() 30 submit->in_fence = NULL; in etnaviv_sched_dependency() 38 for (i = 0; i < submit->nr_bos; i++) { in etnaviv_sched_dependency() 39 struct etnaviv_gem_submit_bo *bo = &submit->bos[i]; in etnaviv_sched_dependency() 74 struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); in etnaviv_sched_run_job() local 78 fence = etnaviv_gpu_submit(submit); in etnaviv_sched_run_job() 80 dev_dbg(submit->gpu->dev, "skipping bad job\n"); in etnaviv_sched_run_job() 87 struct etnaviv_gem_submit *submit = to_etnaviv_submit(sched_job); in etnaviv_sched_timedout_job() local [all …]
|
D | etnaviv_dump.c | 116 struct etnaviv_gem_submit *submit; in etnaviv_core_dump() local 139 submit = to_etnaviv_submit(s_job); in etnaviv_core_dump() 140 file_size += submit->cmdbuf.size; in etnaviv_core_dump() 187 submit = to_etnaviv_submit(s_job); in etnaviv_core_dump() 189 submit->cmdbuf.vaddr, submit->cmdbuf.size, in etnaviv_core_dump() 190 etnaviv_cmdbuf_get_va(&submit->cmdbuf)); in etnaviv_core_dump()
|
D | etnaviv_gpu.c | 1212 const struct etnaviv_gem_submit *submit = event->submit; in sync_point_perfmon_sample() local 1215 for (i = 0; i < submit->nr_pmrs; i++) { in sync_point_perfmon_sample() 1216 const struct etnaviv_perfmon_request *pmr = submit->pmrs + i; in sync_point_perfmon_sample() 1219 etnaviv_perfmon_process(gpu, pmr, submit->exec_state); in sync_point_perfmon_sample() 1244 const struct etnaviv_gem_submit *submit = event->submit; in sync_point_perfmon_sample_post() local 1250 for (i = 0; i < submit->nr_pmrs; i++) { in sync_point_perfmon_sample_post() 1251 const struct etnaviv_perfmon_request *pmr = submit->pmrs + i; in sync_point_perfmon_sample_post() 1269 struct dma_fence *etnaviv_gpu_submit(struct etnaviv_gem_submit *submit) in etnaviv_gpu_submit() argument 1271 struct etnaviv_gpu *gpu = submit->gpu; in etnaviv_gpu_submit() 1276 if (!submit->runtime_resumed) { in etnaviv_gpu_submit() [all …]
|
D | etnaviv_gpu.h | 84 struct etnaviv_gem_submit *submit; member 184 struct dma_fence *etnaviv_gpu_submit(struct etnaviv_gem_submit *submit);
|
D | etnaviv_sched.h | 22 struct etnaviv_gem_submit *submit);
|
/Linux-v4.19/crypto/async_tx/ |
D | async_raid6_recov.c | 33 size_t len, struct async_submit_ctl *submit) in async_sum_product() argument 35 struct dma_chan *chan = async_tx_find_channel(submit, DMA_PQ, in async_sum_product() 52 if (submit->flags & ASYNC_TX_FENCE) in async_sum_product() 68 async_tx_submit(chan, tx, submit); in async_sum_product() 80 async_tx_quiesce(&submit->depend_tx); in async_sum_product() 98 struct async_submit_ctl *submit) in async_mult() argument 100 struct dma_chan *chan = async_tx_find_channel(submit, DMA_PQ, in async_mult() 116 if (submit->flags & ASYNC_TX_FENCE) in async_mult() 135 async_tx_submit(chan, tx, submit); in async_mult() 148 async_tx_quiesce(&submit->depend_tx); in async_mult() [all …]
|
D | async_xor.c | 37 struct async_submit_ctl *submit) in do_async_xor() argument 41 dma_async_tx_callback cb_fn_orig = submit->cb_fn; in do_async_xor() 42 void *cb_param_orig = submit->cb_param; in do_async_xor() 43 enum async_tx_flags flags_orig = submit->flags; in do_async_xor() 53 submit->flags = flags_orig; in do_async_xor() 59 submit->flags &= ~ASYNC_TX_ACK; in do_async_xor() 60 submit->flags |= ASYNC_TX_FENCE; in do_async_xor() 61 submit->cb_fn = NULL; in do_async_xor() 62 submit->cb_param = NULL; in do_async_xor() 64 submit->cb_fn = cb_fn_orig; in do_async_xor() [all …]
|
D | async_pq.c | 55 struct async_submit_ctl *submit) in do_async_gen_syndrome() argument 59 enum async_tx_flags flags_orig = submit->flags; in do_async_gen_syndrome() 60 dma_async_tx_callback cb_fn_orig = submit->cb_fn; in do_async_gen_syndrome() 61 dma_async_tx_callback cb_param_orig = submit->cb_param; in do_async_gen_syndrome() 68 submit->flags = flags_orig; in do_async_gen_syndrome() 75 submit->flags &= ~ASYNC_TX_ACK; in do_async_gen_syndrome() 76 submit->flags |= ASYNC_TX_FENCE; in do_async_gen_syndrome() 77 submit->cb_fn = NULL; in do_async_gen_syndrome() 78 submit->cb_param = NULL; in do_async_gen_syndrome() 80 submit->cb_fn = cb_fn_orig; in do_async_gen_syndrome() [all …]
|
D | raid6test.c | 70 struct async_submit_ctl submit; in raid6_dual_recov() local 81 init_async_submit(&submit, 0, NULL, NULL, NULL, addr_conv); in raid6_dual_recov() 82 tx = async_gen_syndrome(ptrs, 0, disks, bytes, &submit); in raid6_dual_recov() 100 init_async_submit(&submit, ASYNC_TX_XOR_ZERO_DST, NULL, in raid6_dual_recov() 102 tx = async_xor(dest, blocks, 0, count, bytes, &submit); in raid6_dual_recov() 104 init_async_submit(&submit, 0, tx, NULL, NULL, addr_conv); in raid6_dual_recov() 105 tx = async_gen_syndrome(ptrs, 0, disks, bytes, &submit); in raid6_dual_recov() 110 init_async_submit(&submit, 0, NULL, NULL, NULL, addr_conv); in raid6_dual_recov() 111 tx = async_raid6_datap_recov(disks, bytes, faila, ptrs, &submit); in raid6_dual_recov() 114 init_async_submit(&submit, 0, NULL, NULL, NULL, addr_conv); in raid6_dual_recov() [all …]
|
D | async_tx.c | 56 __async_tx_find_channel(struct async_submit_ctl *submit, in __async_tx_find_channel() argument 59 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in __async_tx_find_channel() 157 struct async_submit_ctl *submit) in async_tx_submit() argument 159 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in async_tx_submit() 161 tx->callback = submit->cb_fn; in async_tx_submit() 162 tx->callback_param = submit->cb_param; in async_tx_submit() 217 if (submit->flags & ASYNC_TX_ACK) in async_tx_submit() 234 async_trigger_callback(struct async_submit_ctl *submit) in async_trigger_callback() argument 239 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in async_trigger_callback() 258 async_tx_submit(chan, tx, submit); in async_trigger_callback() [all …]
|
D | async_memcpy.c | 47 struct async_submit_ctl *submit) in async_memcpy() argument 49 struct dma_chan *chan = async_tx_find_channel(submit, DMA_MEMCPY, in async_memcpy() 61 if (submit->cb_fn) in async_memcpy() 63 if (submit->flags & ASYNC_TX_FENCE) in async_memcpy() 83 async_tx_submit(chan, tx, submit); in async_memcpy() 89 async_tx_quiesce(&submit->depend_tx); in async_memcpy() 99 async_tx_sync_epilog(submit); in async_memcpy()
|
/Linux-v4.19/include/linux/ |
D | async_tx.h | 116 __async_tx_find_channel(struct async_submit_ctl *submit, 131 async_tx_find_channel(struct async_submit_ctl *submit, in async_tx_find_channel() argument 146 async_tx_sync_epilog(struct async_submit_ctl *submit) in async_tx_sync_epilog() argument 148 if (submit->cb_fn) in async_tx_sync_epilog() 149 submit->cb_fn(submit->cb_param); in async_tx_sync_epilog() 172 struct async_submit_ctl *submit); 176 int src_cnt, size_t len, struct async_submit_ctl *submit); 181 struct async_submit_ctl *submit); 186 struct async_submit_ctl *submit); 188 struct dma_async_tx_descriptor *async_trigger_callback(struct async_submit_ctl *submit); [all …]
|
/Linux-v4.19/drivers/dma/ti/ |
D | cppi41.c | 118 u16 submit; member 156 [ 0] = { .submit = 32, .complete = 93}, 157 [ 1] = { .submit = 34, .complete = 94}, 158 [ 2] = { .submit = 36, .complete = 95}, 159 [ 3] = { .submit = 38, .complete = 96}, 160 [ 4] = { .submit = 40, .complete = 97}, 161 [ 5] = { .submit = 42, .complete = 98}, 162 [ 6] = { .submit = 44, .complete = 99}, 163 [ 7] = { .submit = 46, .complete = 100}, 164 [ 8] = { .submit = 48, .complete = 101}, [all …]
|
/Linux-v4.19/drivers/gpu/drm/msm/adreno/ |
D | a5xx_gpu.c | 147 static void a5xx_submit_in_rb(struct msm_gpu *gpu, struct msm_gem_submit *submit, in a5xx_submit_in_rb() argument 151 struct msm_ringbuffer *ring = submit->ring; in a5xx_submit_in_rb() 156 for (i = 0; i < submit->nr_cmds; i++) { in a5xx_submit_in_rb() 157 switch (submit->cmd[i].type) { in a5xx_submit_in_rb() 165 obj = submit->bos[submit->cmd[i].idx].obj; in a5xx_submit_in_rb() 166 dwords = submit->cmd[i].size; in a5xx_submit_in_rb() 202 ring->memptrs->fence = submit->seqno; in a5xx_submit_in_rb() 206 static void a5xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit, in a5xx_submit() argument 212 struct msm_ringbuffer *ring = submit->ring; in a5xx_submit() 215 if (IS_ENABLED(CONFIG_DRM_MSM_GPU_SUDO) && submit->in_rb) { in a5xx_submit() [all …]
|
D | a6xx_gpu.c | 68 static void a6xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit, in a6xx_submit() argument 72 struct msm_ringbuffer *ring = submit->ring; in a6xx_submit() 83 for (i = 0; i < submit->nr_cmds; i++) { in a6xx_submit() 84 switch (submit->cmd[i].type) { in a6xx_submit() 92 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova)); in a6xx_submit() 93 OUT_RING(ring, upper_32_bits(submit->cmd[i].iova)); in a6xx_submit() 94 OUT_RING(ring, submit->cmd[i].size); in a6xx_submit() 101 OUT_RING(ring, submit->seqno); in a6xx_submit() 111 OUT_RING(ring, submit->seqno); in a6xx_submit() 763 .submit = a6xx_submit,
|
D | adreno_gpu.c | 285 void adreno_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit, in adreno_submit() argument 290 struct msm_ringbuffer *ring = submit->ring; in adreno_submit() 293 for (i = 0; i < submit->nr_cmds; i++) { in adreno_submit() 294 switch (submit->cmd[i].type) { in adreno_submit() 305 OUT_RING(ring, lower_32_bits(submit->cmd[i].iova)); in adreno_submit() 306 OUT_RING(ring, submit->cmd[i].size); in adreno_submit() 313 OUT_RING(ring, submit->seqno); in adreno_submit() 331 OUT_RING(ring, submit->seqno); in adreno_submit()
|
/Linux-v4.19/Documentation/crypto/ |
D | async-tx-api.txt | 57 async_<operation>(<op specific parameters>, struct async_submit ctl *submit) 79 operations complete. When an application needs to submit a chain of 110 async_<operation>, or when the application needs to submit a chain of 117 2/ Completion callback routines cannot submit new operations. This 142 struct async_submit_ctl submit; 146 init_async_submit(&submit, ASYNC_TX_XOR_DROP_DST, NULL, NULL, NULL, 148 tx = async_xor(xor_dest, xor_srcs, 0, xor_src_cnt, xor_len, &submit) 150 submit->depend_tx = tx; 151 tx = async_memcpy(copy_dest, copy_src, 0, 0, copy_len, &submit); 154 init_async_submit(&submit, ASYNC_TX_XOR_DROP_DST | ASYNC_TX_ACK, tx, [all …]
|
/Linux-v4.19/drivers/gpu/drm/i915/ |
D | i915_request.c | 83 i915_sw_fence_fini(&rq->submit); in i915_fence_release() 420 GEM_BUG_ON(!i915_sw_fence_signaled(&request->submit)); in i915_request_retire() 622 container_of(fence, typeof(*request), submit); in submit_notify() 779 i915_sw_fence_init(&i915_request_get(rq)->submit, submit_notify); in i915_request_alloc() 864 ret = i915_sw_fence_await_sw_fence_gfp(&to->submit, in i915_request_await_request() 865 &from->submit, in i915_request_await_request() 892 ret = i915_sw_fence_await_dma_fence(&to->submit, in i915_request_await_request() 942 ret = i915_sw_fence_await_dma_fence(&rq->submit, fence, in i915_request_await_dma_fence() 1093 i915_sw_fence_await_sw_fence(&request->submit, &prev->submit, in i915_request_add() 1132 i915_sw_fence_commit(&request->submit); in i915_request_add() [all …]
|
/Linux-v4.19/fs/ |
D | iomap.c | 1431 } submit; member 1527 struct task_struct *waiter = dio->submit.waiter; in iomap_dio_bio_end_io() 1528 WRITE_ONCE(dio->submit.waiter, NULL); in iomap_dio_bio_end_io() 1579 unsigned int align = iov_iter_alignment(dio->submit.iter); in iomap_dio_bio_actor() 1617 iter = *dio->submit.iter; in iomap_dio_bio_actor() 1634 iov_iter_revert(dio->submit.iter, copied); in iomap_dio_bio_actor() 1666 iov_iter_advance(dio->submit.iter, n); in iomap_dio_bio_actor() 1676 dio->submit.last_queue = bdev_get_queue(iomap->bdev); in iomap_dio_bio_actor() 1677 dio->submit.cookie = submit_bio(bio); in iomap_dio_bio_actor() 1692 length = iov_iter_zero(length, dio->submit.iter); in iomap_dio_hole_actor() [all …]
|
/Linux-v4.19/Documentation/translations/ja_JP/ |
D | SubmitChecklist | 2 This is a version of Documentation/process/submit-checklist.rst into Japanese. 17 linux-2.6.26/Documentation/process/submit-checklist.rst の和訳です。
|