Lines Matching refs:workload

501 	struct intel_vgpu_workload *workload;  member
888 u32 base = s->workload->engine->mmio_base; in is_cmd_update_pdps()
896 struct intel_vgpu_mm *shadow_mm = s->workload->shadow_mm; in cmd_pdp_mmio_update_handler()
912 &s->workload->lri_shadow_mm); in cmd_pdp_mmio_update_handler()
982 s->workload->ring_context_gpa + 12, &ctx_sr_ctl, 4); in cmd_reg_handler()
1193 s->workload->pending_events); in cmd_handler_pipe_control()
1200 s->workload->pending_events); in cmd_handler_mi_user_interrupt()
1711 s->workload->pending_events); in cmd_handler_mi_flush_dw()
1782 s->vgpu->gtt.ggtt_mm : s->workload->shadow_mm; in find_bb_size()
1798 s->engine->name, s->workload); in find_bb_size()
1810 s->engine->name, s->workload); in find_bb_size()
1844 s->engine->name, s->workload); in audit_bb_end()
1865 s->vgpu->gtt.ggtt_mm : s->workload->shadow_mm; in perform_bb_shadow()
1926 list_add(&bb->list, &s->workload->shadow_bb); in perform_bb_shadow()
2704 s->engine->name, s->workload); in cmd_parser_exec()
2712 s->workload, info->name); in cmd_parser_exec()
2796 static int scan_workload(struct intel_vgpu_workload *workload) in scan_workload() argument
2803 if (WARN_ON(!IS_ALIGNED(workload->rb_start, I915_GTT_PAGE_SIZE))) in scan_workload()
2806 gma_head = workload->rb_start + workload->rb_head; in scan_workload()
2807 gma_tail = workload->rb_start + workload->rb_tail; in scan_workload()
2808 gma_bottom = workload->rb_start + _RING_CTL_BUF_SIZE(workload->rb_ctl); in scan_workload()
2812 s.vgpu = workload->vgpu; in scan_workload()
2813 s.engine = workload->engine; in scan_workload()
2814 s.ring_start = workload->rb_start; in scan_workload()
2815 s.ring_size = _RING_CTL_BUF_SIZE(workload->rb_ctl); in scan_workload()
2818 s.rb_va = workload->shadow_ring_buffer_va; in scan_workload()
2819 s.workload = workload; in scan_workload()
2822 if (bypass_scan_mask & workload->engine->mask || gma_head == gma_tail) in scan_workload()
2829 ret = command_scan(&s, workload->rb_head, workload->rb_tail, in scan_workload()
2830 workload->rb_start, _RING_CTL_BUF_SIZE(workload->rb_ctl)); in scan_workload()
2842 struct intel_vgpu_workload *workload = container_of(wa_ctx, in scan_wa_ctx() local
2860 s.vgpu = workload->vgpu; in scan_wa_ctx()
2861 s.engine = workload->engine; in scan_wa_ctx()
2867 s.workload = workload; in scan_wa_ctx()
2880 static int shadow_workload_ring_buffer(struct intel_vgpu_workload *workload) in shadow_workload_ring_buffer() argument
2882 struct intel_vgpu *vgpu = workload->vgpu; in shadow_workload_ring_buffer()
2888 guest_rb_size = _RING_CTL_BUF_SIZE(workload->rb_ctl); in shadow_workload_ring_buffer()
2891 workload->rb_len = (workload->rb_tail + guest_rb_size - in shadow_workload_ring_buffer()
2892 workload->rb_head) % guest_rb_size; in shadow_workload_ring_buffer()
2894 gma_head = workload->rb_start + workload->rb_head; in shadow_workload_ring_buffer()
2895 gma_tail = workload->rb_start + workload->rb_tail; in shadow_workload_ring_buffer()
2896 gma_top = workload->rb_start + guest_rb_size; in shadow_workload_ring_buffer()
2898 if (workload->rb_len > s->ring_scan_buffer_size[workload->engine->id]) { in shadow_workload_ring_buffer()
2902 p = krealloc(s->ring_scan_buffer[workload->engine->id], in shadow_workload_ring_buffer()
2903 workload->rb_len, GFP_KERNEL); in shadow_workload_ring_buffer()
2908 s->ring_scan_buffer[workload->engine->id] = p; in shadow_workload_ring_buffer()
2909 s->ring_scan_buffer_size[workload->engine->id] = workload->rb_len; in shadow_workload_ring_buffer()
2912 shadow_ring_buffer_va = s->ring_scan_buffer[workload->engine->id]; in shadow_workload_ring_buffer()
2915 workload->shadow_ring_buffer_va = shadow_ring_buffer_va; in shadow_workload_ring_buffer()
2926 gma_head = workload->rb_start; in shadow_workload_ring_buffer()
2939 int intel_gvt_scan_and_shadow_ringbuffer(struct intel_vgpu_workload *workload) in intel_gvt_scan_and_shadow_ringbuffer() argument
2942 struct intel_vgpu *vgpu = workload->vgpu; in intel_gvt_scan_and_shadow_ringbuffer()
2944 ret = shadow_workload_ring_buffer(workload); in intel_gvt_scan_and_shadow_ringbuffer()
2950 ret = scan_workload(workload); in intel_gvt_scan_and_shadow_ringbuffer()
2962 struct intel_vgpu_workload *workload = container_of(wa_ctx, in shadow_indirect_ctx() local
2965 struct intel_vgpu *vgpu = workload->vgpu; in shadow_indirect_ctx()
2970 obj = i915_gem_object_create_shmem(workload->engine->i915, in shadow_indirect_ctx()
2992 ret = copy_gma_to_hva(workload->vgpu, in shadow_indirect_ctx()
2993 workload->vgpu->gtt.ggtt_mm, in shadow_indirect_ctx()
3034 struct intel_vgpu_workload *workload = container_of(wa_ctx, in intel_gvt_scan_and_shadow_wa_ctx() local
3037 struct intel_vgpu *vgpu = workload->vgpu; in intel_gvt_scan_and_shadow_wa_ctx()