Lines Matching refs:ring

122 int amdgpu_ib_schedule(struct amdgpu_ring *ring, unsigned num_ibs,  in amdgpu_ib_schedule()  argument
126 struct amdgpu_device *adev = ring->adev; in amdgpu_ib_schedule()
153 if (!ring->sched.ready) { in amdgpu_ib_schedule()
154 dev_err(adev->dev, "couldn't schedule ib on ring <%s>\n", ring->name); in amdgpu_ib_schedule()
163 alloc_size = ring->funcs->emit_frame_size + num_ibs * in amdgpu_ib_schedule()
164 ring->funcs->emit_ib_size; in amdgpu_ib_schedule()
166 r = amdgpu_ring_alloc(ring, alloc_size); in amdgpu_ib_schedule()
172 need_ctx_switch = ring->current_ctx != fence_ctx; in amdgpu_ib_schedule()
173 if (ring->funcs->emit_pipeline_sync && job && in amdgpu_ib_schedule()
176 amdgpu_vm_need_pipeline_sync(ring, job))) { in amdgpu_ib_schedule()
185 if (ring->funcs->insert_start) in amdgpu_ib_schedule()
186 ring->funcs->insert_start(ring); in amdgpu_ib_schedule()
189 r = amdgpu_vm_flush(ring, job, need_pipe_sync); in amdgpu_ib_schedule()
191 amdgpu_ring_undo(ring); in amdgpu_ib_schedule()
196 if (job && ring->funcs->init_cond_exec) in amdgpu_ib_schedule()
197 patch_offset = amdgpu_ring_init_cond_exec(ring); in amdgpu_ib_schedule()
203 if (ring->funcs->emit_hdp_flush) in amdgpu_ib_schedule()
204 amdgpu_ring_emit_hdp_flush(ring); in amdgpu_ib_schedule()
206 amdgpu_asic_flush_hdp(adev, ring); in amdgpu_ib_schedule()
212 skip_preamble = ring->current_ctx == fence_ctx; in amdgpu_ib_schedule()
213 if (job && ring->funcs->emit_cntxcntl) { in amdgpu_ib_schedule()
216 amdgpu_ring_emit_cntxcntl(ring, status); in amdgpu_ib_schedule()
230 amdgpu_ring_emit_ib(ring, job, ib, status); in amdgpu_ib_schedule()
234 if (ring->funcs->emit_tmz) in amdgpu_ib_schedule()
235 amdgpu_ring_emit_tmz(ring, false); in amdgpu_ib_schedule()
240 amdgpu_asic_invalidate_hdp(adev, ring); in amdgpu_ib_schedule()
247 amdgpu_ring_emit_fence(ring, job->uf_addr, job->uf_sequence, in amdgpu_ib_schedule()
251 r = amdgpu_fence_emit(ring, f, fence_flags); in amdgpu_ib_schedule()
255 amdgpu_vmid_reset(adev, ring->funcs->vmhub, job->vmid); in amdgpu_ib_schedule()
256 amdgpu_ring_undo(ring); in amdgpu_ib_schedule()
260 if (ring->funcs->insert_end) in amdgpu_ib_schedule()
261 ring->funcs->insert_end(ring); in amdgpu_ib_schedule()
263 if (patch_offset != ~0 && ring->funcs->patch_cond_exec) in amdgpu_ib_schedule()
264 amdgpu_ring_patch_cond_exec(ring, patch_offset); in amdgpu_ib_schedule()
266 ring->current_ctx = fence_ctx; in amdgpu_ib_schedule()
267 if (vm && ring->funcs->emit_switch_buffer) in amdgpu_ib_schedule()
268 amdgpu_ring_emit_switch_buffer(ring); in amdgpu_ib_schedule()
269 amdgpu_ring_commit(ring); in amdgpu_ib_schedule()
358 struct amdgpu_ring *ring = adev->rings[i]; in amdgpu_ib_ring_tests() local
364 if (!ring->sched.ready || !ring->funcs->test_ib) in amdgpu_ib_ring_tests()
368 if (ring->funcs->type == AMDGPU_RING_TYPE_UVD || in amdgpu_ib_ring_tests()
369 ring->funcs->type == AMDGPU_RING_TYPE_VCE || in amdgpu_ib_ring_tests()
370 ring->funcs->type == AMDGPU_RING_TYPE_UVD_ENC || in amdgpu_ib_ring_tests()
371 ring->funcs->type == AMDGPU_RING_TYPE_VCN_DEC || in amdgpu_ib_ring_tests()
372 ring->funcs->type == AMDGPU_RING_TYPE_VCN_ENC || in amdgpu_ib_ring_tests()
373 ring->funcs->type == AMDGPU_RING_TYPE_VCN_JPEG) in amdgpu_ib_ring_tests()
378 r = amdgpu_ring_test_ib(ring, tmo); in amdgpu_ib_ring_tests()
381 ring->name); in amdgpu_ib_ring_tests()
385 ring->sched.ready = false; in amdgpu_ib_ring_tests()
387 ring->name, r); in amdgpu_ib_ring_tests()
389 if (ring == &adev->gfx.gfx_ring[0]) { in amdgpu_ib_ring_tests()