Lines Matching refs:gpu
80 struct etnaviv_gpu *gpu) in etnaviv_core_dump_registers() argument
87 reg->value = gpu_read(gpu, etnaviv_dump_registers[i]); in etnaviv_core_dump_registers()
94 struct etnaviv_gpu *gpu, size_t mmu_size) in etnaviv_core_dump_mmu() argument
96 etnaviv_iommu_dump(gpu->mmu, iter->data); in etnaviv_core_dump_mmu()
111 void etnaviv_core_dump(struct etnaviv_gpu *gpu) in etnaviv_core_dump() argument
127 mmu_size = etnaviv_iommu_dump_size(gpu->mmu); in etnaviv_core_dump()
134 mmu_size + gpu->buffer.size; in etnaviv_core_dump()
137 spin_lock(&gpu->sched.job_list_lock); in etnaviv_core_dump()
138 list_for_each_entry(s_job, &gpu->sched.ring_mirror_list, node) { in etnaviv_core_dump()
143 spin_unlock(&gpu->sched.job_list_lock); in etnaviv_core_dump()
146 list_for_each_entry(vram, &gpu->mmu->mappings, mmu_node) { in etnaviv_core_dump()
169 dev_warn(gpu->dev, "failed to allocate devcoredump file\n"); in etnaviv_core_dump()
179 etnaviv_core_dump_registers(&iter, gpu); in etnaviv_core_dump()
180 etnaviv_core_dump_mmu(&iter, gpu, mmu_size); in etnaviv_core_dump()
181 etnaviv_core_dump_mem(&iter, ETDUMP_BUF_RING, gpu->buffer.vaddr, in etnaviv_core_dump()
182 gpu->buffer.size, in etnaviv_core_dump()
183 etnaviv_cmdbuf_get_va(&gpu->buffer)); in etnaviv_core_dump()
185 spin_lock(&gpu->sched.job_list_lock); in etnaviv_core_dump()
186 list_for_each_entry(s_job, &gpu->sched.ring_mirror_list, node) { in etnaviv_core_dump()
192 spin_unlock(&gpu->sched.job_list_lock); in etnaviv_core_dump()
205 list_for_each_entry(vram, &gpu->mmu->mappings, mmu_node) { in etnaviv_core_dump()
238 dev_coredumpv(gpu->dev, iter.start, iter.data - iter.start, GFP_KERNEL); in etnaviv_core_dump()