Lines Matching refs:tbo
230 return drm_vma_node_verify_access(&abo->tbo.base.vma_node, in amdgpu_verify_access()
789 struct ttm_tt *ttm = bo->tbo.ttm; in amdgpu_ttm_tt_get_user_pages()
998 struct ttm_buffer_object *tbo, in amdgpu_ttm_gart_bind() argument
1001 struct amdgpu_bo *abo = ttm_to_amdgpu_bo(tbo); in amdgpu_ttm_gart_bind()
1002 struct ttm_tt *ttm = tbo->ttm; in amdgpu_ttm_gart_bind()
1151 int amdgpu_ttm_recover_gart(struct ttm_buffer_object *tbo) in amdgpu_ttm_recover_gart() argument
1153 struct amdgpu_device *adev = amdgpu_ttm_adev(tbo->bdev); in amdgpu_ttm_recover_gart()
1157 if (!tbo->ttm) in amdgpu_ttm_recover_gart()
1160 flags = amdgpu_ttm_tt_pte_flags(adev, tbo->ttm, &tbo->mem); in amdgpu_ttm_recover_gart()
1161 r = amdgpu_ttm_gart_bind(adev, tbo, flags); in amdgpu_ttm_recover_gart()
1539 struct amdgpu_device *adev = amdgpu_ttm_adev(abo->tbo.bdev); in amdgpu_ttm_access_memory()
1549 nodes = amdgpu_find_mm_node(&abo->tbo.mem, &offset); in amdgpu_ttm_access_memory()
1679 ttm_bo_mem_put(&bo->tbo, &bo->tbo.mem); in amdgpu_ttm_fw_reserve_vram_init()
1680 r = ttm_bo_mem_space(&bo->tbo, &bo->placement, in amdgpu_ttm_fw_reserve_vram_init()
1681 &bo->tbo.mem, &ctx); in amdgpu_ttm_fw_reserve_vram_init()
2076 struct amdgpu_device *adev = amdgpu_ttm_adev(bo->tbo.bdev); in amdgpu_fill_buffer()
2092 if (bo->tbo.mem.mem_type == TTM_PL_TT) { in amdgpu_fill_buffer()
2093 r = amdgpu_ttm_alloc_gart(&bo->tbo); in amdgpu_fill_buffer()
2098 num_pages = bo->tbo.num_pages; in amdgpu_fill_buffer()
2099 mm_node = bo->tbo.mem.mm_node; in amdgpu_fill_buffer()
2126 num_pages = bo->tbo.num_pages; in amdgpu_fill_buffer()
2127 mm_node = bo->tbo.mem.mm_node; in amdgpu_fill_buffer()
2133 dst_addr = amdgpu_mm_node_addr(&bo->tbo, mm_node, &bo->tbo.mem); in amdgpu_fill_buffer()