Lines Matching refs:tbo

212 	struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev);  in amdgpu_amdkfd_unreserve_memory_limit()
237 struct dma_resv *resv = bo->tbo.base.resv; in amdgpu_amdkfd_remove_eviction_fence()
315 BUG_ON(!dma_resv_trylock(bo->tbo.base.resv)); in amdgpu_amdkfd_remove_fence_on_pt_pd_bos()
317 dma_resv_unlock(bo->tbo.base.resv); in amdgpu_amdkfd_remove_fence_on_pt_pd_bos()
329 if (WARN(amdgpu_ttm_tt_get_usermm(bo->tbo.ttm), in amdgpu_amdkfd_bo_validate()
335 ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in amdgpu_amdkfd_bo_validate()
360 struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); in vm_validate_pt_pd_bos()
391 struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); in vm_update_pds()
403 struct amdgpu_device *bo_adev = amdgpu_ttm_adev(mem->bo->tbo.bdev); in get_pte_flags()
475 struct ttm_tt *src_ttm = mem->bo->tbo.ttm; in kfd_mem_dmamap_userptr()
476 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmamap_userptr()
502 ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in kfd_mem_dmamap_userptr()
526 return ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in kfd_mem_dmamap_dmabuf()
556 struct ttm_tt *ttm = bo->tbo.ttm; in kfd_mem_dmaunmap_userptr()
562 ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in kfd_mem_dmaunmap_userptr()
577 ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in kfd_mem_dmaunmap_dmabuf()
602 unsigned long bo_size = mem->bo->tbo.base.size; in kfd_mem_attach_userptr()
613 ttm_bo_type_sg, mem->bo->tbo.base.resv, in kfd_mem_attach_userptr()
633 mem->dmabuf = amdgpu_gem_prime_export(&mem->bo->tbo.base, in kfd_mem_attach_dmabuf()
676 struct amdgpu_device *bo_adev = amdgpu_ttm_adev(mem->bo->tbo.bdev); in kfd_mem_attach()
677 unsigned long bo_size = mem->bo->tbo.base.size; in kfd_mem_attach()
705 drm_gem_object_get(&bo[i]->tbo.base); in kfd_mem_attach()
710 drm_gem_object_get(&bo[i]->tbo.base); in kfd_mem_attach()
711 } else if (amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm)) { in kfd_mem_attach()
718 mem->bo->tbo.type != ttm_bo_type_sg) { in kfd_mem_attach()
733 drm_gem_object_get(&bo[i]->tbo.base); in kfd_mem_attach()
764 drm_gem_object_put(&bo[i]->tbo.base); in kfd_mem_attach()
777 drm_gem_object_put(&bo->tbo.base); in kfd_mem_detach()
791 entry->bo = &bo->tbo; in add_kgd_mem_to_kfd_bo_list()
832 ret = amdgpu_ttm_tt_set_userptr(&bo->tbo, user_addr, 0); in init_user_pages()
845 ret = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages); in init_user_pages()
857 ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in init_user_pages()
863 amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm); in init_user_pages()
920 ctx->kfd_bo.tv.bo = &bo->tbo; in reserve_bo_and_vm()
983 ctx->kfd_bo.tv.bo = &bo->tbo; in reserve_bo_and_cond_vms()
1159 ret = amdgpu_sync_resv(NULL, sync, pd->tbo.base.resv, in process_sync_pds_resv()
1236 ret = dma_resv_reserve_shared(vm->root.bo->tbo.base.resv, 1); in init_kfd_vm()
1378 struct amdgpu_device *adev = amdgpu_ttm_adev(pd->tbo.bdev); in amdgpu_amdkfd_gpuvm_get_process_page_dir()
1477 bo->tbo.sg = sg; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
1478 bo->tbo.ttm->sg = sg; in amdgpu_amdkfd_gpuvm_alloc_memory_of_gpu()
1527 unsigned long bo_size = mem->bo->tbo.base.size; in amdgpu_amdkfd_gpuvm_free_memory_of_gpu()
1583 if (mem->bo->tbo.sg) { in amdgpu_amdkfd_gpuvm_free_memory_of_gpu()
1584 sg_free_table(mem->bo->tbo.sg); in amdgpu_amdkfd_gpuvm_free_memory_of_gpu()
1585 kfree(mem->bo->tbo.sg); in amdgpu_amdkfd_gpuvm_free_memory_of_gpu()
1600 drm_vma_node_revoke(&mem->bo->tbo.base.vma_node, drm_priv); in amdgpu_amdkfd_gpuvm_free_memory_of_gpu()
1603 drm_gem_object_put(&mem->bo->tbo.base); in amdgpu_amdkfd_gpuvm_free_memory_of_gpu()
1640 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
1649 bo_size = bo->tbo.base.size; in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
1671 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) && in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
1672 bo->tbo.resource->mem_type == TTM_PL_SYSTEM) in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
1680 !amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
1718 if (!amdgpu_ttm_tt_get_usermm(bo->tbo.ttm) && !bo->tbo.pin_count) in amdgpu_amdkfd_gpuvm_map_memory_to_gpu()
1745 unsigned long bo_size = mem->bo->tbo.base.size; in amdgpu_amdkfd_gpuvm_unmap_memory_from_gpu()
1789 !amdgpu_ttm_tt_get_usermm(mem->bo->tbo.ttm) && in amdgpu_amdkfd_gpuvm_unmap_memory_from_gpu()
1790 !mem->bo->tbo.pin_count) in amdgpu_amdkfd_gpuvm_unmap_memory_from_gpu()
1824 if (amdgpu_ttm_tt_get_usermm(bo->tbo.ttm)) { in amdgpu_amdkfd_gpuvm_map_gtt_bo_to_kernel()
1940 drm_gem_object_get(&bo->tbo.base); in amdgpu_amdkfd_gpuvm_import_dmabuf()
2014 ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in update_invalid_user_pages()
2042 ret = amdgpu_ttm_tt_get_user_pages(bo, bo->tbo.ttm->pages); in update_invalid_user_pages()
2055 amdgpu_ttm_tt_get_user_pages_done(bo->tbo.ttm); in update_invalid_user_pages()
2133 if (bo->tbo.ttm->pages[0]) { in validate_invalid_user_pages()
2135 ret = ttm_bo_validate(&bo->tbo, &bo->placement, &ctx); in validate_invalid_user_pages()
2352 ret = amdgpu_sync_fence(&sync_obj, bo->tbo.moving); in amdgpu_amdkfd_gpuvm_restore_process_bos()
2462 ret = dma_resv_reserve_shared(gws_bo->tbo.base.resv, 1); in amdgpu_amdkfd_add_gws_to_process()