/Linux-v4.19/drivers/gpu/drm/radeon/ |
D | radeon_gem.c | 34 struct radeon_bo *robj = gem_to_radeon_bo(gobj); in radeon_gem_object_free() local 36 if (robj) { in radeon_gem_object_free() 37 radeon_mn_unregister(robj); in radeon_gem_object_free() 38 radeon_bo_unref(&robj); in radeon_gem_object_free() 47 struct radeon_bo *robj; in radeon_gem_object_create() local 69 flags, NULL, NULL, &robj); in radeon_gem_object_create() 81 *obj = &robj->gem_base; in radeon_gem_object_create() 82 robj->pid = task_pid_nr(current); in radeon_gem_object_create() 85 list_add_tail(&robj->list, &rdev->gem.objects); in radeon_gem_object_create() 94 struct radeon_bo *robj; in radeon_gem_set_domain() local [all …]
|
D | radeon_gart.c | 131 if (rdev->gart.robj == NULL) { in radeon_gart_table_vram_alloc() 134 0, NULL, NULL, &rdev->gart.robj); in radeon_gart_table_vram_alloc() 157 r = radeon_bo_reserve(rdev->gart.robj, false); in radeon_gart_table_vram_pin() 160 r = radeon_bo_pin(rdev->gart.robj, in radeon_gart_table_vram_pin() 163 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_pin() 166 r = radeon_bo_kmap(rdev->gart.robj, &rdev->gart.ptr); in radeon_gart_table_vram_pin() 168 radeon_bo_unpin(rdev->gart.robj); in radeon_gart_table_vram_pin() 169 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_pin() 199 if (rdev->gart.robj == NULL) { in radeon_gart_table_vram_unpin() 202 r = radeon_bo_reserve(rdev->gart.robj, false); in radeon_gart_table_vram_unpin() [all …]
|
D | r100_track.h | 14 struct radeon_bo *robj; member 21 struct radeon_bo *robj; member 26 struct radeon_bo *robj; member 37 struct radeon_bo *robj; member
|
D | evergreen_cs.c | 1222 track->db_z_read_bo = reloc->robj; in evergreen_cs_handle_reg() 1234 track->db_z_write_bo = reloc->robj; in evergreen_cs_handle_reg() 1246 track->db_s_read_bo = reloc->robj; in evergreen_cs_handle_reg() 1258 track->db_s_write_bo = reloc->robj; in evergreen_cs_handle_reg() 1282 track->vgt_strmout_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1505 track->cb_color_fmask_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1522 track->cb_color_cmask_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1563 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1579 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg() 1591 track->htile_bo = reloc->robj; in evergreen_cs_handle_reg() [all …]
|
D | radeon_cursor.c | 285 struct radeon_bo *robj; in radeon_crtc_cursor_set2() local 307 robj = gem_to_radeon_bo(obj); in radeon_crtc_cursor_set2() 308 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2() 314 ret = radeon_bo_pin_restricted(robj, RADEON_GEM_DOMAIN_VRAM, in radeon_crtc_cursor_set2() 317 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2() 349 struct radeon_bo *robj = gem_to_radeon_bo(radeon_crtc->cursor_bo); in radeon_crtc_cursor_set2() local 350 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2() 352 radeon_bo_unpin(robj); in radeon_crtc_cursor_set2() 353 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2()
|
D | r600_cs.c | 1086 track->vgt_strmout_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1213 track->cb_color_frag_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1244 track->cb_color_tile_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1284 track->cb_color_bo[tmp] = reloc->robj; in r600_cs_check_reg() 1297 track->db_bo = reloc->robj; in r600_cs_check_reg() 1310 track->htile_bo = reloc->robj; in r600_cs_check_reg() 1808 if ((tmp + size) > radeon_bo_size(reloc->robj)) { in r600_packet3_check() 1810 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check() 1838 if ((tmp + size) > radeon_bo_size(reloc->robj)) { in r600_packet3_check() 1840 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check() [all …]
|
D | r200.c | 188 track->zb.robj = reloc->robj; in r200_packet0_check() 201 track->cb[0].robj = reloc->robj; in r200_packet0_check() 231 track->textures[i].robj = reloc->robj; in r200_packet0_check() 275 track->textures[i].cube_info[face - 1].robj = reloc->robj; in r200_packet0_check()
|
D | r100.c | 1328 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr() 1338 track->arrays[i + 1].robj = reloc->robj; in r100_packet3_load_vbpntr() 1352 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr() 1589 track->zb.robj = reloc->robj; in r100_packet0_check() 1602 track->cb[0].robj = reloc->robj; in r100_packet0_check() 1629 track->textures[i].robj = reloc->robj; in r100_packet0_check() 1647 track->textures[0].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1665 track->textures[1].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1683 track->textures[2].cube_info[i].robj = reloc->robj; in r100_packet0_check() 1888 struct radeon_bo *robj) in r100_cs_track_check_pkt3_indx_buffer() argument [all …]
|
D | radeon_device.c | 1595 struct radeon_bo *robj; in radeon_suspend_kms() local 1598 struct radeon_bo *robj = gem_to_radeon_bo(radeon_crtc->cursor_bo); in radeon_suspend_kms() local 1599 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms() 1601 radeon_bo_unpin(robj); in radeon_suspend_kms() 1602 radeon_bo_unreserve(robj); in radeon_suspend_kms() 1609 robj = gem_to_radeon_bo(fb->obj[0]); in radeon_suspend_kms() 1611 if (!radeon_fbdev_robj_is_fb(rdev, robj)) { in radeon_suspend_kms() 1612 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms() 1614 radeon_bo_unpin(robj); in radeon_suspend_kms() 1615 radeon_bo_unreserve(robj); in radeon_suspend_kms() [all …]
|
D | radeon_cs.c | 110 p->relocs[i].robj = gem_to_radeon_bo(gobj); in radeon_cs_parser_relocs() 157 if (radeon_ttm_tt_has_userptr(p->relocs[i].robj->tbo.ttm)) { in radeon_cs_parser_relocs() 171 if (p->relocs[i].robj->prime_shared_count) { in radeon_cs_parser_relocs() 180 p->relocs[i].tv.bo = &p->relocs[i].robj->tbo; in radeon_cs_parser_relocs() 254 resv = reloc->robj->tbo.resv; in radeon_cs_sync_rings() 398 return (int)la->robj->tbo.num_pages - (int)lb->robj->tbo.num_pages; in cmp_size_smaller_first() 436 struct radeon_bo *bo = parser->relocs[i].robj; in radeon_cs_parser_fini() 519 bo = p->relocs[i].robj; in radeon_bo_vm_update_pte()
|
D | r300.c | 129 if (rdev->gart.robj) { in rv370_pcie_gart_init() 153 if (rdev->gart.robj == NULL) { in rv370_pcie_gart_enable() 670 track->cb[i].robj = reloc->robj; in r300_packet0_check() 683 track->zb.robj = reloc->robj; in r300_packet0_check() 728 track->textures[i].robj = reloc->robj; in r300_packet0_check() 1127 track->aa.robj = reloc->robj; in r300_packet0_check() 1196 r = r100_cs_track_check_pkt3_indx_buffer(p, pkt, reloc->robj); in r300_packet3_check()
|
D | radeon_fb.c | 398 bool radeon_fbdev_robj_is_fb(struct radeon_device *rdev, struct radeon_bo *robj) in radeon_fbdev_robj_is_fb() argument 403 if (robj == gem_to_radeon_bo(rdev->mode_info.rfbdev->fb.obj[0])) in radeon_fbdev_robj_is_fb()
|
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_gart.c | 115 if (adev->gart.robj == NULL) { in amdgpu_gart_table_vram_alloc() 126 r = amdgpu_bo_create(adev, &bp, &adev->gart.robj); in amdgpu_gart_table_vram_alloc() 148 r = amdgpu_bo_reserve(adev->gart.robj, false); in amdgpu_gart_table_vram_pin() 151 r = amdgpu_bo_pin(adev->gart.robj, AMDGPU_GEM_DOMAIN_VRAM); in amdgpu_gart_table_vram_pin() 153 amdgpu_bo_unreserve(adev->gart.robj); in amdgpu_gart_table_vram_pin() 156 r = amdgpu_bo_kmap(adev->gart.robj, &adev->gart.ptr); in amdgpu_gart_table_vram_pin() 158 amdgpu_bo_unpin(adev->gart.robj); in amdgpu_gart_table_vram_pin() 159 amdgpu_bo_unreserve(adev->gart.robj); in amdgpu_gart_table_vram_pin() 160 adev->gart.table_addr = amdgpu_bo_gpu_offset(adev->gart.robj); in amdgpu_gart_table_vram_pin() 176 if (adev->gart.robj == NULL) { in amdgpu_gart_table_vram_unpin() [all …]
|
D | amdgpu_gem.c | 37 struct amdgpu_bo *robj = gem_to_amdgpu_bo(gobj); in amdgpu_gem_object_free() local 39 if (robj) { in amdgpu_gem_object_free() 40 amdgpu_mn_unregister(robj); in amdgpu_gem_object_free() 41 amdgpu_bo_unref(&robj); in amdgpu_gem_object_free() 377 struct amdgpu_bo *robj; in amdgpu_mode_dumb_mmap() local 383 robj = gem_to_amdgpu_bo(gobj); in amdgpu_mode_dumb_mmap() 384 if (amdgpu_ttm_tt_get_usermm(robj->tbo.ttm) || in amdgpu_mode_dumb_mmap() 385 (robj->flags & AMDGPU_GEM_CREATE_NO_CPU_ACCESS)) { in amdgpu_mode_dumb_mmap() 389 *offset_p = amdgpu_bo_mmap_offset(robj); in amdgpu_mode_dumb_mmap() 436 struct amdgpu_bo *robj; in amdgpu_gem_wait_idle_ioctl() local [all …]
|
D | amdgpu_bo_list.c | 53 amdgpu_bo_unref(&e->robj); in amdgpu_bo_list_free() 114 entry->robj = bo; in amdgpu_bo_list_create() 117 entry->tv.bo = &entry->robj->tbo; in amdgpu_bo_list_create() 118 entry->tv.shared = !entry->robj->prime_shared_count; in amdgpu_bo_list_create() 120 if (entry->robj->preferred_domains == AMDGPU_GEM_DOMAIN_GDS) in amdgpu_bo_list_create() 121 list->gds_obj = entry->robj; in amdgpu_bo_list_create() 122 if (entry->robj->preferred_domains == AMDGPU_GEM_DOMAIN_GWS) in amdgpu_bo_list_create() 123 list->gws_obj = entry->robj; in amdgpu_bo_list_create() 124 if (entry->robj->preferred_domains == AMDGPU_GEM_DOMAIN_OA) in amdgpu_bo_list_create() 125 list->oa_obj = entry->robj; in amdgpu_bo_list_create() [all …]
|
D | amdgpu_cs.c | 48 p->uf_entry.robj = amdgpu_bo_ref(gem_to_amdgpu_bo(gobj)); in amdgpu_cs_user_fence_chunk() 50 p->uf_entry.tv.bo = &p->uf_entry.robj->tbo; in amdgpu_cs_user_fence_chunk() 56 size = amdgpu_bo_size(p->uf_entry.robj); in amdgpu_cs_user_fence_chunk() 62 if (amdgpu_ttm_tt_get_usermm(p->uf_entry.robj->tbo.ttm)) { in amdgpu_cs_user_fence_chunk() 72 amdgpu_bo_unref(&p->uf_entry.robj); in amdgpu_cs_user_fence_chunk() 231 if (p->uf_entry.robj) in amdgpu_cs_parser_init() 460 struct amdgpu_bo *bo = candidate->robj; in amdgpu_cs_try_evict() 466 if (candidate->robj == validated) in amdgpu_cs_try_evict() 531 struct amdgpu_bo *bo = lobj->robj; in amdgpu_cs_list_validate() 606 if (p->uf_entry.robj && !p->uf_entry.robj->parent) in amdgpu_cs_parser_bos() [all …]
|
D | amdgpu_fb.c | 382 struct amdgpu_bo *robj; in amdgpu_fbdev_total_size() local 388 robj = gem_to_amdgpu_bo(adev->mode_info.rfbdev->rfb.base.obj[0]); in amdgpu_fbdev_total_size() 389 size += amdgpu_bo_size(robj); in amdgpu_fbdev_total_size() 393 bool amdgpu_fbdev_robj_is_fb(struct amdgpu_device *adev, struct amdgpu_bo *robj) in amdgpu_fbdev_robj_is_fb() argument 397 if (robj == gem_to_amdgpu_bo(adev->mode_info.rfbdev->rfb.base.obj[0])) in amdgpu_fbdev_robj_is_fb()
|
D | amdgpu_gart.h | 44 struct amdgpu_bo *robj; member
|
D | amdgpu_bo_list.h | 35 struct amdgpu_bo *robj; member
|
/Linux-v4.19/drivers/gpu/drm/nouveau/ |
D | nouveau_prime.c | 65 struct reservation_object *robj = attach->dmabuf->resv; in nouveau_gem_prime_import_sg_table() local 71 ww_mutex_lock(&robj->lock, NULL); in nouveau_gem_prime_import_sg_table() 73 sg, robj, &nvbo); in nouveau_gem_prime_import_sg_table() 74 ww_mutex_unlock(&robj->lock); in nouveau_gem_prime_import_sg_table()
|
D | nouveau_bo.h | 79 struct reservation_object *robj,
|
/Linux-v4.19/drivers/gpu/drm/etnaviv/ |
D | etnaviv_gem.c | 462 struct reservation_object *robj = etnaviv_obj->resv; in etnaviv_gem_describe() local 473 fobj = rcu_dereference(robj->fence); in etnaviv_gem_describe() 483 fence = rcu_dereference(robj->fence_excl); in etnaviv_gem_describe() 570 struct reservation_object *robj, const struct etnaviv_gem_ops *ops, in etnaviv_gem_new_impl() argument 599 if (robj) { in etnaviv_gem_new_impl() 600 etnaviv_obj->resv = robj; in etnaviv_gem_new_impl() 661 struct reservation_object *robj, const struct etnaviv_gem_ops *ops, in etnaviv_gem_new_private() argument 667 ret = etnaviv_gem_new_impl(dev, size, flags, robj, ops, &obj); in etnaviv_gem_new_private()
|
D | etnaviv_gem_submit.c | 179 struct reservation_object *robj = bo->obj->resv; in submit_fence_sync() local 182 ret = reservation_object_reserve_shared(robj); in submit_fence_sync() 191 ret = reservation_object_get_fences_rcu(robj, &bo->excl, in submit_fence_sync() 197 bo->excl = reservation_object_get_excl_rcu(robj); in submit_fence_sync()
|
D | etnaviv_gem.h | 118 struct reservation_object *robj, const struct etnaviv_gem_ops *ops,
|
/Linux-v4.19/drivers/dma-buf/ |
D | dma-buf.c | 1016 struct reservation_object *robj; in dma_buf_debug_show() local 1047 robj = buf_obj->resv; in dma_buf_debug_show() 1049 seq = read_seqcount_begin(&robj->seq); in dma_buf_debug_show() 1051 fobj = rcu_dereference(robj->fence); in dma_buf_debug_show() 1053 fence = rcu_dereference(robj->fence_excl); in dma_buf_debug_show() 1054 if (!read_seqcount_retry(&robj->seq, seq)) in dma_buf_debug_show()
|