Home
last modified time | relevance | path

Searched refs:wa_ctx (Results 1 – 9 of 9) sorted by relevance

/Linux-v5.10/drivers/gpu/drm/i915/gvt/
Dscheduler.c394 static void release_shadow_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx) in release_shadow_wa_ctx() argument
396 if (!wa_ctx->indirect_ctx.obj) in release_shadow_wa_ctx()
399 i915_gem_object_unpin_map(wa_ctx->indirect_ctx.obj); in release_shadow_wa_ctx()
400 i915_gem_object_put(wa_ctx->indirect_ctx.obj); in release_shadow_wa_ctx()
402 wa_ctx->indirect_ctx.obj = NULL; in release_shadow_wa_ctx()
403 wa_ctx->indirect_ctx.shadow_va = NULL; in release_shadow_wa_ctx()
486 workload->wa_ctx.indirect_ctx.size) { in intel_gvt_scan_and_shadow_workload()
487 ret = intel_gvt_scan_and_shadow_wa_ctx(&workload->wa_ctx); in intel_gvt_scan_and_shadow_workload()
496 release_shadow_wa_ctx(&workload->wa_ctx); in intel_gvt_scan_and_shadow_workload()
559 static void update_wa_ctx_2_shadow_ctx(struct intel_shadow_wa_ctx *wa_ctx) in update_wa_ctx_2_shadow_ctx() argument
[all …]
Dcmd_parser.c2836 static int scan_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx) in scan_wa_ctx() argument
2842 struct intel_vgpu_workload *workload = container_of(wa_ctx, in scan_wa_ctx()
2844 wa_ctx); in scan_wa_ctx()
2847 if (WARN_ON(!IS_ALIGNED(wa_ctx->indirect_ctx.guest_gma, in scan_wa_ctx()
2851 ring_tail = wa_ctx->indirect_ctx.size + 3 * sizeof(u32); in scan_wa_ctx()
2852 ring_size = round_up(wa_ctx->indirect_ctx.size + CACHELINE_BYTES, in scan_wa_ctx()
2854 gma_head = wa_ctx->indirect_ctx.guest_gma; in scan_wa_ctx()
2855 gma_tail = wa_ctx->indirect_ctx.guest_gma + ring_tail; in scan_wa_ctx()
2856 gma_bottom = wa_ctx->indirect_ctx.guest_gma + ring_size; in scan_wa_ctx()
2862 s.ring_start = wa_ctx->indirect_ctx.guest_gma; in scan_wa_ctx()
[all …]
Dcmd_parser.h51 int intel_gvt_scan_and_shadow_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx);
Dscheduler.h114 struct intel_shadow_wa_ctx wa_ctx; member
/Linux-v5.10/drivers/gpu/drm/i915/gt/
Dselftest_ring_submission.c223 engine->wa_ctx.vma = bb; in __live_ctx_switch_wa()
238 intel_context_put(engine->wa_ctx.vma->private); in __live_ctx_switch_wa()
239 i915_vma_unpin_and_release(&engine->wa_ctx.vma, I915_VMA_RELEASE_MAP); in __live_ctx_switch_wa()
272 saved_wa = fetch_and_zero(&engine->wa_ctx.vma); in live_ctx_switch_wa()
280 engine->wa_ctx.vma = saved_wa; in live_ctx_switch_wa()
Dintel_ring_submission.c865 engine->wa_ctx.vma->node.start, 0, in clear_residuals()
887 if (engine->wa_ctx.vma && ce != engine->kernel_context) { in switch_context()
888 if (engine->wa_ctx.vma->private != ce) { in switch_context()
893 residuals = &engine->wa_ctx.vma->private; in switch_context()
1030 if (engine->wa_ctx.vma) { in ring_release()
1031 intel_context_put(engine->wa_ctx.vma->private); in ring_release()
1032 i915_vma_unpin_and_release(&engine->wa_ctx.vma, 0); in ring_release()
1226 engine->wa_ctx.vma = vma; in gen7_ctx_switch_bb_init()
Dintel_lrc.c3485 GEM_BUG_ON(engine->wa_ctx.indirect_ctx.size); in __execlists_update_reg_state()
3928 engine->wa_ctx.vma = vma; in lrc_setup_wa_ctx()
3938 i915_vma_unpin_and_release(&engine->wa_ctx.vma, 0); in lrc_destroy_wa_ctx()
3945 struct i915_ctx_workarounds *wa_ctx = &engine->wa_ctx; in intel_init_workaround_bb() local
3946 struct i915_wa_ctx_bb *wa_bb[2] = { &wa_ctx->indirect_ctx, in intel_init_workaround_bb()
3947 &wa_ctx->per_ctx }; in intel_init_workaround_bb()
3984 batch = i915_gem_object_pin_map(wa_ctx->vma->obj, I915_MAP_WB); in intel_init_workaround_bb()
4005 __i915_gem_object_flush_map(wa_ctx->vma->obj, 0, batch_ptr - batch); in intel_init_workaround_bb()
4006 __i915_gem_object_release_map(wa_ctx->vma->obj); in intel_init_workaround_bb()
5240 const struct i915_ctx_workarounds * const wa_ctx = &engine->wa_ctx; in init_wa_bb_reg_state() local
[all …]
Dintel_engine_types.h405 struct i915_ctx_workarounds wa_ctx; member
/Linux-v5.10/drivers/gpu/drm/i915/
Di915_gpu_error.c1418 engine->wa_ctx.vma, in intel_engine_coredump_add_vma()