Lines Matching refs:batch

229 	struct i915_vma *batch; /** identity of the batch obj/vma */  member
541 eb->batch = vma; in eb_add_vma()
744 unsigned int i, batch; in eb_lookup_vmas() local
753 batch = eb_batch_index(eb); in eb_lookup_vmas()
805 err = eb_add_vma(eb, i, batch, vma); in eb_lookup_vmas()
933 GEM_BUG_ON(cache->rq_size >= cache->rq->batch->obj->base.size / sizeof(u32)); in reloc_gpu_flush()
936 __i915_gem_object_flush_map(cache->rq->batch->obj, 0, cache->rq_size); in reloc_gpu_flush()
937 i915_gem_object_unpin_map(cache->rq->batch->obj); in reloc_gpu_flush()
1146 struct i915_vma *batch; in __reloc_gpu_alloc() local
1163 batch = i915_vma_instance(pool->obj, vma->vm, NULL); in __reloc_gpu_alloc()
1164 if (IS_ERR(batch)) { in __reloc_gpu_alloc()
1165 err = PTR_ERR(batch); in __reloc_gpu_alloc()
1169 err = i915_vma_pin(batch, 0, 0, PIN_USER | PIN_NONBLOCK); in __reloc_gpu_alloc()
1188 batch->node.start, PAGE_SIZE, in __reloc_gpu_alloc()
1193 i915_vma_lock(batch); in __reloc_gpu_alloc()
1194 err = i915_request_await_object(rq, batch->obj, false); in __reloc_gpu_alloc()
1196 err = i915_vma_move_to_active(batch, rq, 0); in __reloc_gpu_alloc()
1197 i915_vma_unlock(batch); in __reloc_gpu_alloc()
1201 rq->batch = batch; in __reloc_gpu_alloc()
1202 i915_vma_unpin(batch); in __reloc_gpu_alloc()
1216 i915_vma_unpin(batch); in __reloc_gpu_alloc()
1271 u32 *batch; in relocate_entry() local
1281 batch = reloc_gpu(eb, vma, len); in relocate_entry()
1282 if (IS_ERR(batch)) in relocate_entry()
1288 *batch++ = MI_STORE_DWORD_IMM_GEN4; in relocate_entry()
1289 *batch++ = lower_32_bits(addr); in relocate_entry()
1290 *batch++ = upper_32_bits(addr); in relocate_entry()
1291 *batch++ = lower_32_bits(target_offset); in relocate_entry()
1295 *batch++ = MI_STORE_DWORD_IMM_GEN4; in relocate_entry()
1296 *batch++ = lower_32_bits(addr); in relocate_entry()
1297 *batch++ = upper_32_bits(addr); in relocate_entry()
1298 *batch++ = upper_32_bits(target_offset); in relocate_entry()
1300 *batch++ = (MI_STORE_DWORD_IMM_GEN4 | (1 << 21)) + 1; in relocate_entry()
1301 *batch++ = lower_32_bits(addr); in relocate_entry()
1302 *batch++ = upper_32_bits(addr); in relocate_entry()
1303 *batch++ = lower_32_bits(target_offset); in relocate_entry()
1304 *batch++ = upper_32_bits(target_offset); in relocate_entry()
1307 *batch++ = MI_STORE_DWORD_IMM_GEN4; in relocate_entry()
1308 *batch++ = 0; in relocate_entry()
1309 *batch++ = addr; in relocate_entry()
1310 *batch++ = target_offset; in relocate_entry()
1312 *batch++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT; in relocate_entry()
1313 *batch++ = 0; in relocate_entry()
1314 *batch++ = addr; in relocate_entry()
1315 *batch++ = target_offset; in relocate_entry()
1317 *batch++ = MI_STORE_DWORD_IMM | MI_MEM_VIRTUAL; in relocate_entry()
1318 *batch++ = addr; in relocate_entry()
1319 *batch++ = target_offset; in relocate_entry()
1735 GEM_BUG_ON(!eb->batch); in eb_relocate_slow()
2003 batch_start = gen8_canonical_addr(eb->batch->node.start) + in eb_parse()
2010 eb->batch->obj, in eb_parse()
2041 eb->batch = vma; in eb_parse()
2095 eb->batch->node.start + in eb_submit()
2585 if (unlikely(*eb.batch->exec_flags & EXEC_OBJECT_WRITE)) { in i915_gem_do_execbuffer()
2590 if (eb.batch_start_offset > eb.batch->size || in i915_gem_do_execbuffer()
2591 eb.batch_len > eb.batch->size - eb.batch_start_offset) { in i915_gem_do_execbuffer()
2598 eb.batch_len = eb.batch->size - eb.batch_start_offset; in i915_gem_do_execbuffer()
2627 vma = i915_gem_object_ggtt_pin(eb.batch->obj, NULL, 0, 0, 0); in i915_gem_do_execbuffer()
2633 eb.batch = vma; in i915_gem_do_execbuffer()
2680 eb.request->batch = eb.batch; in i915_gem_do_execbuffer()
2681 if (eb.batch->private) in i915_gem_do_execbuffer()
2682 intel_engine_pool_mark_active(eb.batch->private, eb.request); in i915_gem_do_execbuffer()
2706 i915_vma_unpin(eb.batch); in i915_gem_do_execbuffer()
2707 if (eb.batch->private) in i915_gem_do_execbuffer()
2708 intel_engine_pool_put(eb.batch->private); in i915_gem_do_execbuffer()