/Linux-v4.19/drivers/gpu/drm/radeon/ |
D | radeon_vm.c | 296 struct radeon_bo_va *bo_va; in radeon_vm_bo_find() local 298 list_for_each_entry(bo_va, &bo->va, bo_list) { in radeon_vm_bo_find() 299 if (bo_va->vm == vm) { in radeon_vm_bo_find() 300 return bo_va; in radeon_vm_bo_find() 323 struct radeon_bo_va *bo_va; in radeon_vm_bo_add() local 325 bo_va = kzalloc(sizeof(struct radeon_bo_va), GFP_KERNEL); in radeon_vm_bo_add() 326 if (bo_va == NULL) { in radeon_vm_bo_add() 329 bo_va->vm = vm; in radeon_vm_bo_add() 330 bo_va->bo = bo; in radeon_vm_bo_add() 331 bo_va->it.start = 0; in radeon_vm_bo_add() [all …]
|
D | radeon_gem.c | 149 struct radeon_bo_va *bo_va; in radeon_gem_object_open() local 162 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_open() 163 if (!bo_va) { in radeon_gem_object_open() 164 bo_va = radeon_vm_bo_add(rdev, vm, rbo); in radeon_gem_object_open() 166 ++bo_va->ref_count; in radeon_gem_object_open() 180 struct radeon_bo_va *bo_va; in radeon_gem_object_close() local 194 bo_va = radeon_vm_bo_find(vm, rbo); in radeon_gem_object_close() 195 if (bo_va) { in radeon_gem_object_close() 196 if (--bo_va->ref_count == 0) { in radeon_gem_object_close() 197 radeon_vm_bo_rmv(rdev, bo_va); in radeon_gem_object_close() [all …]
|
D | radeon_trace.h | 66 TP_PROTO(struct radeon_bo_va *bo_va), 67 TP_ARGS(bo_va), 75 __entry->soffset = bo_va->it.start; 76 __entry->eoffset = bo_va->it.last + 1; 77 __entry->flags = bo_va->flags;
|
D | radeon_cs.c | 495 struct radeon_bo_va *bo_va; in radeon_bo_vm_update_pte() local 520 bo_va = radeon_vm_bo_find(vm, bo); in radeon_bo_vm_update_pte() 521 if (bo_va == NULL) { in radeon_bo_vm_update_pte() 526 r = radeon_vm_bo_update(rdev, bo_va, &bo->tbo.mem); in radeon_bo_vm_update_pte() 530 radeon_sync_fence(&p->ib.sync, bo_va->last_pt_update); in radeon_bo_vm_update_pte()
|
D | radeon.h | 2861 struct radeon_bo_va *bo_va, 2871 struct radeon_bo_va *bo_va, 2875 struct radeon_bo_va *bo_va);
|
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_vm.c | 802 struct amdgpu_bo_va *bo_va; in amdgpu_vm_bo_find() local 804 list_for_each_entry(bo_va, &bo->va, base.bo_list) { in amdgpu_vm_bo_find() 805 if (bo_va->base.vm == vm) { in amdgpu_vm_bo_find() 806 return bo_va; in amdgpu_vm_bo_find() 1638 struct amdgpu_bo_va *bo_va, in amdgpu_vm_bo_update() argument 1641 struct amdgpu_bo *bo = bo_va->base.bo; in amdgpu_vm_bo_update() 1642 struct amdgpu_vm *vm = bo_va->base.vm; in amdgpu_vm_bo_update() 1675 last_update = &bo_va->last_pt_update; in amdgpu_vm_bo_update() 1677 if (!clear && bo_va->base.moved) { in amdgpu_vm_bo_update() 1678 bo_va->base.moved = false; in amdgpu_vm_bo_update() [all …]
|
D | amdgpu_gem.c | 127 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_open() local 143 bo_va = amdgpu_vm_bo_find(vm, abo); in amdgpu_gem_object_open() 144 if (!bo_va) { in amdgpu_gem_object_open() 145 bo_va = amdgpu_vm_bo_add(adev, vm, abo); in amdgpu_gem_object_open() 147 ++bo_va->ref_count; in amdgpu_gem_object_open() 165 struct amdgpu_bo_va *bo_va; in amdgpu_gem_object_close() local 183 bo_va = amdgpu_vm_bo_find(vm, bo); in amdgpu_gem_object_close() 184 if (bo_va && --bo_va->ref_count == 0) { in amdgpu_gem_object_close() 185 amdgpu_vm_bo_rmv(adev, bo_va); in amdgpu_gem_object_close() 520 struct amdgpu_bo_va *bo_va, in amdgpu_gem_va_update_vm() argument [all …]
|
D | amdgpu_vm.h | 296 struct amdgpu_bo_va *bo_va, 306 struct amdgpu_bo_va *bo_va, 310 struct amdgpu_bo_va *bo_va, 314 struct amdgpu_bo_va *bo_va, 323 struct amdgpu_bo_va *bo_va);
|
D | amdgpu_virt.c | 77 struct amdgpu_bo_va **bo_va) in amdgpu_map_static_csa() argument 100 *bo_va = amdgpu_vm_bo_add(adev, vm, adev->virt.csa_obj); in amdgpu_map_static_csa() 101 if (!*bo_va) { in amdgpu_map_static_csa() 107 r = amdgpu_vm_alloc_pts(adev, (*bo_va)->base.vm, csa_addr, in amdgpu_map_static_csa() 111 amdgpu_vm_bo_rmv(adev, *bo_va); in amdgpu_map_static_csa() 116 r = amdgpu_vm_bo_map(adev, *bo_va, csa_addr, 0, AMDGPU_CSA_SIZE, in amdgpu_map_static_csa() 122 amdgpu_vm_bo_rmv(adev, *bo_va); in amdgpu_map_static_csa()
|
D | amdgpu_amdkfd_gpuvm.c | 86 if (entry->bo_va->base.vm == avm) in check_if_add_bo_to_vm() 471 bo_va_entry->bo_va = amdgpu_vm_bo_add(adev, vm, bo); in add_bo_to_vm() 472 if (!bo_va_entry->bo_va) { in add_bo_to_vm() 516 amdgpu_vm_bo_rmv(adev, bo_va_entry->bo_va); in add_bo_to_vm() 529 amdgpu_vm_bo_rmv(adev, entry->bo_va); in remove_bo_from_vm() 731 if ((vm && vm != entry->bo_va->base.vm) || in reserve_bo_and_cond_vms() 755 if ((vm && vm != entry->bo_va->base.vm) || in reserve_bo_and_cond_vms() 760 amdgpu_vm_get_pd_bo(entry->bo_va->base.vm, &ctx->list, in reserve_bo_and_cond_vms() 814 struct amdgpu_bo_va *bo_va = entry->bo_va; in unmap_bo_from_gpuvm() local 815 struct amdgpu_vm *vm = bo_va->base.vm; in unmap_bo_from_gpuvm() [all …]
|
D | amdgpu_trace.h | 242 TP_PROTO(struct amdgpu_bo_va *bo_va, 244 TP_ARGS(bo_va, mapping), 254 __entry->bo = bo_va ? bo_va->base.bo : NULL; 266 TP_PROTO(struct amdgpu_bo_va *bo_va, 268 TP_ARGS(bo_va, mapping), 278 __entry->bo = bo_va ? bo_va->base.bo : NULL;
|
D | amdgpu_cs.c | 719 e->bo_va = amdgpu_vm_bo_find(vm, e->robj); in amdgpu_cs_parser_bos() 819 struct amdgpu_bo_va *bo_va; in amdgpu_bo_vm_update_pte() local 839 bo_va = fpriv->csa_va; in amdgpu_bo_vm_update_pte() 840 BUG_ON(!bo_va); in amdgpu_bo_vm_update_pte() 841 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_bo_vm_update_pte() 845 f = bo_va->last_pt_update; in amdgpu_bo_vm_update_pte() 859 bo_va = e->bo_va; in amdgpu_bo_vm_update_pte() 860 if (bo_va == NULL) in amdgpu_bo_vm_update_pte() 863 r = amdgpu_vm_bo_update(adev, bo_va, false); in amdgpu_bo_vm_update_pte() 867 f = bo_va->last_pt_update; in amdgpu_bo_vm_update_pte() [all …]
|
D | amdgpu_bo_list.h | 37 struct amdgpu_bo_va *bo_va; member
|
D | amdgpu_amdkfd.h | 43 struct amdgpu_bo_va *bo_va; member
|
D | amdgpu_virt.h | 286 struct amdgpu_bo_va **bo_va);
|
D | amdgpu_object.h | 49 struct amdgpu_bo_va *bo_va; member
|