Lines Matching refs:batch
43 u32 *batch; member
84 h->batch = vaddr; in hang_init()
115 u32 *batch; in emit_recurse_batch() local
152 batch = h->batch; in emit_recurse_batch()
154 *batch++ = MI_STORE_DWORD_IMM_GEN4; in emit_recurse_batch()
155 *batch++ = lower_32_bits(hws_address(hws, rq)); in emit_recurse_batch()
156 *batch++ = upper_32_bits(hws_address(hws, rq)); in emit_recurse_batch()
157 *batch++ = rq->fence.seqno; in emit_recurse_batch()
158 *batch++ = MI_ARB_CHECK; in emit_recurse_batch()
160 memset(batch, 0, 1024); in emit_recurse_batch()
161 batch += 1024 / sizeof(*batch); in emit_recurse_batch()
163 *batch++ = MI_ARB_CHECK; in emit_recurse_batch()
164 *batch++ = MI_BATCH_BUFFER_START | 1 << 8 | 1; in emit_recurse_batch()
165 *batch++ = lower_32_bits(vma->node.start); in emit_recurse_batch()
166 *batch++ = upper_32_bits(vma->node.start); in emit_recurse_batch()
168 *batch++ = MI_STORE_DWORD_IMM_GEN4; in emit_recurse_batch()
169 *batch++ = 0; in emit_recurse_batch()
170 *batch++ = lower_32_bits(hws_address(hws, rq)); in emit_recurse_batch()
171 *batch++ = rq->fence.seqno; in emit_recurse_batch()
172 *batch++ = MI_ARB_CHECK; in emit_recurse_batch()
174 memset(batch, 0, 1024); in emit_recurse_batch()
175 batch += 1024 / sizeof(*batch); in emit_recurse_batch()
177 *batch++ = MI_ARB_CHECK; in emit_recurse_batch()
178 *batch++ = MI_BATCH_BUFFER_START | 1 << 8; in emit_recurse_batch()
179 *batch++ = lower_32_bits(vma->node.start); in emit_recurse_batch()
181 *batch++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT; in emit_recurse_batch()
182 *batch++ = 0; in emit_recurse_batch()
183 *batch++ = lower_32_bits(hws_address(hws, rq)); in emit_recurse_batch()
184 *batch++ = rq->fence.seqno; in emit_recurse_batch()
185 *batch++ = MI_ARB_CHECK; in emit_recurse_batch()
187 memset(batch, 0, 1024); in emit_recurse_batch()
188 batch += 1024 / sizeof(*batch); in emit_recurse_batch()
190 *batch++ = MI_ARB_CHECK; in emit_recurse_batch()
191 *batch++ = MI_BATCH_BUFFER_START | 2 << 6; in emit_recurse_batch()
192 *batch++ = lower_32_bits(vma->node.start); in emit_recurse_batch()
194 *batch++ = MI_STORE_DWORD_IMM | MI_MEM_VIRTUAL; in emit_recurse_batch()
195 *batch++ = lower_32_bits(hws_address(hws, rq)); in emit_recurse_batch()
196 *batch++ = rq->fence.seqno; in emit_recurse_batch()
197 *batch++ = MI_ARB_CHECK; in emit_recurse_batch()
199 memset(batch, 0, 1024); in emit_recurse_batch()
200 batch += 1024 / sizeof(*batch); in emit_recurse_batch()
202 *batch++ = MI_ARB_CHECK; in emit_recurse_batch()
203 *batch++ = MI_BATCH_BUFFER_START | 2 << 6; in emit_recurse_batch()
204 *batch++ = lower_32_bits(vma->node.start); in emit_recurse_batch()
206 *batch++ = MI_BATCH_BUFFER_END; /* not reached */ in emit_recurse_batch()
247 h->batch = vaddr; in hang_create_request()
270 *h->batch = MI_BATCH_BUFFER_END; in hang_fini()
326 *h.batch = MI_BATCH_BUFFER_END; in igt_hang_sanitycheck()
1307 *h.batch = MI_BATCH_BUFFER_END; in igt_reset_queue()