Home
last modified time | relevance | path

Searched refs:bo_va (Results 1 – 16 of 16) sorted by relevance

/Linux-v4.19/drivers/gpu/drm/radeon/
Dradeon_vm.c296 struct radeon_bo_va *bo_va; in radeon_vm_bo_find() local
298 list_for_each_entry(bo_va, &bo->va, bo_list) { in radeon_vm_bo_find()
299 if (bo_va->vm == vm) { in radeon_vm_bo_find()
300 return bo_va; in radeon_vm_bo_find()
323 struct radeon_bo_va *bo_va; in radeon_vm_bo_add() local
325 bo_va = kzalloc(sizeof(struct radeon_bo_va), GFP_KERNEL); in radeon_vm_bo_add()
326 if (bo_va == NULL) { in radeon_vm_bo_add()
329 bo_va->vm = vm; in radeon_vm_bo_add()
330 bo_va->bo = bo; in radeon_vm_bo_add()
331 bo_va->it.start = 0; in radeon_vm_bo_add()
[all …]
Dradeon_gem.c149 struct radeon_bo_va *bo_va; in radeon_gem_object_open() local
162 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_open()
163 if (!bo_va) { in radeon_gem_object_open()
164 bo_va = radeon_vm_bo_add(rdev, vm, rbo); in radeon_gem_object_open()
166 ++bo_va->ref_count; in radeon_gem_object_open()
180 struct radeon_bo_va *bo_va; in radeon_gem_object_close() local
194 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_close()
195 if (bo_va) { in radeon_gem_object_close()
196 if (--bo_va->ref_count == 0) { in radeon_gem_object_close()
197 radeon_vm_bo_rmv(rdev, bo_va); in radeon_gem_object_close()
[all …]
Dradeon_trace.h66 TP_PROTO(struct radeon_bo_va *bo_va),
67 TP_ARGS(bo_va),
75 __entry->soffset = bo_va->it.start;
76 __entry->eoffset = bo_va->it.last + 1;
77 __entry->flags = bo_va->flags;
Dradeon_cs.c495 struct radeon_bo_va *bo_va; in radeon_bo_vm_update_pte() local
520 bo_va = radeon_vm_bo_find(vm, bo); in radeon_bo_vm_update_pte()
521 if (bo_va == NULL) { in radeon_bo_vm_update_pte()
526 r = radeon_vm_bo_update(rdev, bo_va, &bo->tbo.mem); in radeon_bo_vm_update_pte()
530 radeon_sync_fence(&p->ib.sync, bo_va->last_pt_update); in radeon_bo_vm_update_pte()
Dradeon.h2861 struct radeon_bo_va *bo_va,
2871 struct radeon_bo_va *bo_va,
2875 struct radeon_bo_va *bo_va);
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/
Damdgpu_vm.c802 struct amdgpu_bo_va *bo_va; in amdgpu_vm_bo_find() local
804 list_for_each_entry(bo_va, &bo->va, base.bo_list) { in amdgpu_vm_bo_find()
805 if (bo_va->base.vm == vm) { in amdgpu_vm_bo_find()
806 return bo_va; in amdgpu_vm_bo_find()
1638 struct amdgpu_bo_va *bo_va, in amdgpu_vm_bo_update() argument
1641 struct amdgpu_bo *bo = bo_va->base.bo; in amdgpu_vm_bo_update()
1642 struct amdgpu_vm *vm = bo_va->base.vm; in amdgpu_vm_bo_update()
1675 last_update = &bo_va->last_pt_update; in amdgpu_vm_bo_update()
1677 if (!clear && bo_va->base.moved) { in amdgpu_vm_bo_update()
1678 bo_va->base.moved = false; in amdgpu_vm_bo_update()
[all …]
Damdgpu_gem.c127 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_open() local
143 bo_va = amdgpu_vm_bo_find(vm, abo); in amdgpu_gem_object_open()
144 if (!bo_va) { in amdgpu_gem_object_open()
145 bo_va = amdgpu_vm_bo_add(adev, vm, abo); in amdgpu_gem_object_open()
147 ++bo_va->ref_count; in amdgpu_gem_object_open()
165 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_close() local
183 bo_va = amdgpu_vm_bo_find(vm, bo); in amdgpu_gem_object_close()
184 if (bo_va && --bo_va->ref_count == 0) { in amdgpu_gem_object_close()
185 amdgpu_vm_bo_rmv(adev, bo_va); in amdgpu_gem_object_close()
520 struct amdgpu_bo_va *bo_va, in amdgpu_gem_va_update_vm() argument
[all …]
Damdgpu_vm.h296 struct amdgpu_bo_va *bo_va,
306 struct amdgpu_bo_va *bo_va,
310 struct amdgpu_bo_va *bo_va,
314 struct amdgpu_bo_va *bo_va,
323 struct amdgpu_bo_va *bo_va);
Damdgpu_virt.c77 struct amdgpu_bo_va **bo_va) in amdgpu_map_static_csa() argument
100 *bo_va = amdgpu_vm_bo_add(adev, vm, adev->virt.csa_obj); in amdgpu_map_static_csa()
101 if (!*bo_va) { in amdgpu_map_static_csa()
107 r = amdgpu_vm_alloc_pts(adev, (*bo_va)->base.vm, csa_addr, in amdgpu_map_static_csa()
111 amdgpu_vm_bo_rmv(adev, *bo_va); in amdgpu_map_static_csa()
116 r = amdgpu_vm_bo_map(adev, *bo_va, csa_addr, 0, AMDGPU_CSA_SIZE, in amdgpu_map_static_csa()
122 amdgpu_vm_bo_rmv(adev, *bo_va); in amdgpu_map_static_csa()
Damdgpu_amdkfd_gpuvm.c86 if (entry->bo_va->base.vm == avm) in check_if_add_bo_to_vm()
471 bo_va_entry->bo_va = amdgpu_vm_bo_add(adev, vm, bo); in add_bo_to_vm()
472 if (!bo_va_entry->bo_va) { in add_bo_to_vm()
516 amdgpu_vm_bo_rmv(adev, bo_va_entry->bo_va); in add_bo_to_vm()
529 amdgpu_vm_bo_rmv(adev, entry->bo_va); in remove_bo_from_vm()
731 if ((vm && vm != entry->bo_va->base.vm) || in reserve_bo_and_cond_vms()
755 if ((vm && vm != entry->bo_va->base.vm) || in reserve_bo_and_cond_vms()
760 amdgpu_vm_get_pd_bo(entry->bo_va->base.vm, &ctx->list, in reserve_bo_and_cond_vms()
814 struct amdgpu_bo_va *bo_va = entry->bo_va; in unmap_bo_from_gpuvm() local
815 struct amdgpu_vm *vm = bo_va->base.vm; in unmap_bo_from_gpuvm()
[all …]
Damdgpu_trace.h242 TP_PROTO(struct amdgpu_bo_va *bo_va,
244 TP_ARGS(bo_va, mapping),
254 __entry->bo = bo_va ? bo_va->base.bo : NULL;
266 TP_PROTO(struct amdgpu_bo_va *bo_va,
268 TP_ARGS(bo_va, mapping),
278 __entry->bo = bo_va ? bo_va->base.bo : NULL;
Damdgpu_cs.c719 e->bo_va = amdgpu_vm_bo_find(vm, e->robj); in amdgpu_cs_parser_bos()
819 struct amdgpu_bo_va *bo_va; in amdgpu_bo_vm_update_pte() local
839 bo_va = fpriv->csa_va; in amdgpu_bo_vm_update_pte()
840 BUG_ON(!bo_va); in amdgpu_bo_vm_update_pte()
841 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_bo_vm_update_pte()
845 f = bo_va->last_pt_update; in amdgpu_bo_vm_update_pte()
859 bo_va = e->bo_va; in amdgpu_bo_vm_update_pte()
860 if (bo_va == NULL) in amdgpu_bo_vm_update_pte()
863 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_bo_vm_update_pte()
867 f = bo_va->last_pt_update; in amdgpu_bo_vm_update_pte()
[all …]
Damdgpu_bo_list.h37 struct amdgpu_bo_va *bo_va; member
Damdgpu_amdkfd.h43 struct amdgpu_bo_va *bo_va; member
Damdgpu_virt.h286 struct amdgpu_bo_va **bo_va);
Damdgpu_object.h49 struct amdgpu_bo_va *bo_va; member