/Linux-v4.19/drivers/gpu/drm/i915/selftests/ |
D | intel_workarounds.c | 29 cs = i915_gem_object_pin_map(result, I915_MAP_WB); in read_nonprivs() 132 vaddr = i915_gem_object_pin_map(results, I915_MAP_WB); in check_whitelist()
|
D | i915_request.c | 420 cmd = i915_gem_object_pin_map(obj, I915_MAP_WB); in empty_batch() 586 cmd = i915_gem_object_pin_map(obj, I915_MAP_WC); in recursive_batch() 619 cmd = i915_gem_object_pin_map(batch->obj, I915_MAP_WC); in recursive_batch_resolve() 843 cmd = i915_gem_object_pin_map(request[id]->batch->obj, in live_sequential_engines()
|
D | i915_gem_coherency.c | 160 map = i915_gem_object_pin_map(obj, I915_MAP_WC); in wc_set() 181 map = i915_gem_object_pin_map(obj, I915_MAP_WC); in wc_get()
|
D | i915_gem_dmabuf.c | 138 obj_map = i915_gem_object_pin_map(obj, I915_MAP_WB); in igt_dmabuf_import() 311 ptr = i915_gem_object_pin_map(obj, I915_MAP_WB); in igt_dmabuf_export_kmap()
|
D | intel_lrc.c | 44 vaddr = i915_gem_object_pin_map(spin->hws, I915_MAP_WB); in spinner_init() 52 vaddr = i915_gem_object_pin_map(spin->obj, mode); in spinner_init()
|
D | intel_hangcheck.c | 71 vaddr = i915_gem_object_pin_map(h->hws, I915_MAP_WB); in hang_init() 78 vaddr = i915_gem_object_pin_map(h->obj, in hang_init() 236 vaddr = i915_gem_object_pin_map(obj, in hang_create_request()
|
D | i915_gem_context.c | 50 cmd = i915_gem_object_pin_map(obj, I915_MAP_WB); in gpu_fill_dw()
|
D | huge_pages.c | 909 cmd = i915_gem_object_pin_map(obj, I915_MAP_WB); in gpu_write_dw() 1564 vaddr = i915_gem_object_pin_map(obj, I915_MAP_WB); in igt_tmpfs_fallback()
|
/Linux-v4.19/drivers/gpu/drm/i915/ |
D | i915_gem_dmabuf.c | 104 return i915_gem_object_pin_map(obj, I915_MAP_WB); in i915_gem_dmabuf_vmap()
|
D | intel_guc_log.c | 364 vaddr = i915_gem_object_pin_map(log->vma->obj, I915_MAP_WC); in guc_log_map()
|
D | intel_lrc.c | 1329 vaddr = i915_gem_object_pin_map(ce->state->obj, I915_MAP_WB); in __execlists_context_pin() 2675 vaddr = i915_gem_object_pin_map(ctx_obj, I915_MAP_WB); in populate_lr_context() 2692 defaults = i915_gem_object_pin_map(engine->default_state, in populate_lr_context() 2808 reg = i915_gem_object_pin_map(ce->state->obj, in intel_lr_context_resume()
|
D | i915_cmd_parser.c | 1070 dst = i915_gem_object_pin_map(dst_obj, I915_MAP_FORCE_WB); in copy_batch() 1077 src = i915_gem_object_pin_map(src_obj, I915_MAP_WC); in copy_batch()
|
D | intel_guc_submission.c | 318 vaddr = i915_gem_object_pin_map(vma->obj, I915_MAP_WB); in guc_stage_desc_pool_create() 975 vaddr = i915_gem_object_pin_map(vma->obj, I915_MAP_WB); in guc_client_alloc()
|
D | intel_guc.c | 171 vaddr = i915_gem_object_pin_map(vma->obj, I915_MAP_WB); in guc_shared_data_create()
|
D | intel_ringbuffer.c | 1055 addr = i915_gem_object_pin_map(vma->obj, map); in intel_ring_pin() 1282 vaddr = i915_gem_object_pin_map(obj, I915_MAP_WB); in alloc_context_vma() 1288 defaults = i915_gem_object_pin_map(engine->default_state, in alloc_context_vma()
|
D | intel_engine_cs.c | 605 vaddr = i915_gem_object_pin_map(obj, I915_MAP_WB); in init_status_page() 1162 map = i915_gem_object_pin_map(engine->default_state, in intel_engines_unpark()
|
D | intel_guc_ct.c | 189 blob = i915_gem_object_pin_map(vma->obj, I915_MAP_WB); in ctch_init()
|
D | i915_perf.c | 1526 i915_gem_object_pin_map(bo, I915_MAP_WB); in alloc_oa_buffer() 1854 regs = i915_gem_object_pin_map(ce->state->obj, I915_MAP_WB); in gen8_configure_all_contexts()
|
D | i915_gem_execbuffer.c | 1129 cmd = i915_gem_object_pin_map(obj, in __reloc_gpu_alloc()
|
D | i915_drv.h | 3057 void *__must_check i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
|
D | i915_gem.c | 2826 void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj, in i915_gem_object_pin_map() function
|
D | i915_debugfs.c | 2512 log = i915_gem_object_pin_map(obj, I915_MAP_WC); in i915_guc_log_dump()
|
/Linux-v4.19/drivers/gpu/drm/i915/gvt/ |
D | cmd_parser.c | 1749 bb->va = i915_gem_object_pin_map(bb->obj, I915_MAP_WB); in perform_bb_shadow() 2811 map = i915_gem_object_pin_map(obj, I915_MAP_WB); in shadow_indirect_ctx()
|