Home
last modified time | relevance | path

Searched refs:robj (Results 1 – 25 of 50) sorted by relevance

12

/Linux-v4.19/drivers/gpu/drm/radeon/
Dradeon_gem.c34 struct radeon_bo *robj = gem_to_radeon_bo(gobj); in radeon_gem_object_free() local
36 if (robj) { in radeon_gem_object_free()
37 radeon_mn_unregister(robj); in radeon_gem_object_free()
38 radeon_bo_unref(&robj); in radeon_gem_object_free()
47 struct radeon_bo *robj; in radeon_gem_object_create() local
69 flags, NULL, NULL, &robj); in radeon_gem_object_create()
81 *obj = &robj->gem_base; in radeon_gem_object_create()
82 robj->pid = task_pid_nr(current); in radeon_gem_object_create()
85 list_add_tail(&robj->list, &rdev->gem.objects); in radeon_gem_object_create()
94 struct radeon_bo *robj; in radeon_gem_set_domain() local
[all …]
Dradeon_gart.c131 if (rdev->gart.robj == NULL) { in radeon_gart_table_vram_alloc()
134 0, NULL, NULL, &rdev->gart.robj); in radeon_gart_table_vram_alloc()
157 r = radeon_bo_reserve(rdev->gart.robj, false); in radeon_gart_table_vram_pin()
160 r = radeon_bo_pin(rdev->gart.robj, in radeon_gart_table_vram_pin()
163 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_pin()
166 r = radeon_bo_kmap(rdev->gart.robj, &rdev->gart.ptr); in radeon_gart_table_vram_pin()
168 radeon_bo_unpin(rdev->gart.robj); in radeon_gart_table_vram_pin()
169 radeon_bo_unreserve(rdev->gart.robj); in radeon_gart_table_vram_pin()
199 if (rdev->gart.robj == NULL) { in radeon_gart_table_vram_unpin()
202 r = radeon_bo_reserve(rdev->gart.robj, false); in radeon_gart_table_vram_unpin()
[all …]
Dr100_track.h14 struct radeon_bo *robj; member
21 struct radeon_bo *robj; member
26 struct radeon_bo *robj; member
37 struct radeon_bo *robj; member
Devergreen_cs.c1222 track->db_z_read_bo = reloc->robj; in evergreen_cs_handle_reg()
1234 track->db_z_write_bo = reloc->robj; in evergreen_cs_handle_reg()
1246 track->db_s_read_bo = reloc->robj; in evergreen_cs_handle_reg()
1258 track->db_s_write_bo = reloc->robj; in evergreen_cs_handle_reg()
1282 track->vgt_strmout_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg()
1505 track->cb_color_fmask_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg()
1522 track->cb_color_cmask_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg()
1563 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg()
1579 track->cb_color_bo[tmp] = reloc->robj; in evergreen_cs_handle_reg()
1591 track->htile_bo = reloc->robj; in evergreen_cs_handle_reg()
[all …]
Dradeon_cursor.c285 struct radeon_bo *robj; in radeon_crtc_cursor_set2() local
307 robj = gem_to_radeon_bo(obj); in radeon_crtc_cursor_set2()
308 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2()
314 ret = radeon_bo_pin_restricted(robj, RADEON_GEM_DOMAIN_VRAM, in radeon_crtc_cursor_set2()
317 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2()
349 struct radeon_bo *robj = gem_to_radeon_bo(radeon_crtc->cursor_bo); in radeon_crtc_cursor_set2() local
350 ret = radeon_bo_reserve(robj, false); in radeon_crtc_cursor_set2()
352 radeon_bo_unpin(robj); in radeon_crtc_cursor_set2()
353 radeon_bo_unreserve(robj); in radeon_crtc_cursor_set2()
Dr600_cs.c1086 track->vgt_strmout_bo[tmp] = reloc->robj; in r600_cs_check_reg()
1213 track->cb_color_frag_bo[tmp] = reloc->robj; in r600_cs_check_reg()
1244 track->cb_color_tile_bo[tmp] = reloc->robj; in r600_cs_check_reg()
1284 track->cb_color_bo[tmp] = reloc->robj; in r600_cs_check_reg()
1297 track->db_bo = reloc->robj; in r600_cs_check_reg()
1310 track->htile_bo = reloc->robj; in r600_cs_check_reg()
1808 if ((tmp + size) > radeon_bo_size(reloc->robj)) { in r600_packet3_check()
1810 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check()
1838 if ((tmp + size) > radeon_bo_size(reloc->robj)) { in r600_packet3_check()
1840 tmp + size, radeon_bo_size(reloc->robj)); in r600_packet3_check()
[all …]
Dr200.c188 track->zb.robj = reloc->robj; in r200_packet0_check()
201 track->cb[0].robj = reloc->robj; in r200_packet0_check()
231 track->textures[i].robj = reloc->robj; in r200_packet0_check()
275 track->textures[i].cube_info[face - 1].robj = reloc->robj; in r200_packet0_check()
Dr100.c1328 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr()
1338 track->arrays[i + 1].robj = reloc->robj; in r100_packet3_load_vbpntr()
1352 track->arrays[i + 0].robj = reloc->robj; in r100_packet3_load_vbpntr()
1589 track->zb.robj = reloc->robj; in r100_packet0_check()
1602 track->cb[0].robj = reloc->robj; in r100_packet0_check()
1629 track->textures[i].robj = reloc->robj; in r100_packet0_check()
1647 track->textures[0].cube_info[i].robj = reloc->robj; in r100_packet0_check()
1665 track->textures[1].cube_info[i].robj = reloc->robj; in r100_packet0_check()
1683 track->textures[2].cube_info[i].robj = reloc->robj; in r100_packet0_check()
1888 struct radeon_bo *robj) in r100_cs_track_check_pkt3_indx_buffer() argument
[all …]
Dradeon_device.c1595 struct radeon_bo *robj; in radeon_suspend_kms() local
1598 struct radeon_bo *robj = gem_to_radeon_bo(radeon_crtc->cursor_bo); in radeon_suspend_kms() local
1599 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms()
1601 radeon_bo_unpin(robj); in radeon_suspend_kms()
1602 radeon_bo_unreserve(robj); in radeon_suspend_kms()
1609 robj = gem_to_radeon_bo(fb->obj[0]); in radeon_suspend_kms()
1611 if (!radeon_fbdev_robj_is_fb(rdev, robj)) { in radeon_suspend_kms()
1612 r = radeon_bo_reserve(robj, false); in radeon_suspend_kms()
1614 radeon_bo_unpin(robj); in radeon_suspend_kms()
1615 radeon_bo_unreserve(robj); in radeon_suspend_kms()
[all …]
Dradeon_cs.c110 p->relocs[i].robj = gem_to_radeon_bo(gobj); in radeon_cs_parser_relocs()
157 if (radeon_ttm_tt_has_userptr(p->relocs[i].robj->tbo.ttm)) { in radeon_cs_parser_relocs()
171 if (p->relocs[i].robj->prime_shared_count) { in radeon_cs_parser_relocs()
180 p->relocs[i].tv.bo = &p->relocs[i].robj->tbo; in radeon_cs_parser_relocs()
254 resv = reloc->robj->tbo.resv; in radeon_cs_sync_rings()
398 return (int)la->robj->tbo.num_pages - (int)lb->robj->tbo.num_pages; in cmp_size_smaller_first()
436 struct radeon_bo *bo = parser->relocs[i].robj; in radeon_cs_parser_fini()
519 bo = p->relocs[i].robj; in radeon_bo_vm_update_pte()
Dr300.c129 if (rdev->gart.robj) { in rv370_pcie_gart_init()
153 if (rdev->gart.robj == NULL) { in rv370_pcie_gart_enable()
670 track->cb[i].robj = reloc->robj; in r300_packet0_check()
683 track->zb.robj = reloc->robj; in r300_packet0_check()
728 track->textures[i].robj = reloc->robj; in r300_packet0_check()
1127 track->aa.robj = reloc->robj; in r300_packet0_check()
1196 r = r100_cs_track_check_pkt3_indx_buffer(p, pkt, reloc->robj); in r300_packet3_check()
Dradeon_fb.c398 bool radeon_fbdev_robj_is_fb(struct radeon_device *rdev, struct radeon_bo *robj) in radeon_fbdev_robj_is_fb() argument
403 if (robj == gem_to_radeon_bo(rdev->mode_info.rfbdev->fb.obj[0])) in radeon_fbdev_robj_is_fb()
/Linux-v4.19/drivers/gpu/drm/amd/amdgpu/
Damdgpu_gart.c115 if (adev->gart.robj == NULL) { in amdgpu_gart_table_vram_alloc()
126 r = amdgpu_bo_create(adev, &bp, &adev->gart.robj); in amdgpu_gart_table_vram_alloc()
148 r = amdgpu_bo_reserve(adev->gart.robj, false); in amdgpu_gart_table_vram_pin()
151 r = amdgpu_bo_pin(adev->gart.robj, AMDGPU_GEM_DOMAIN_VRAM); in amdgpu_gart_table_vram_pin()
153 amdgpu_bo_unreserve(adev->gart.robj); in amdgpu_gart_table_vram_pin()
156 r = amdgpu_bo_kmap(adev->gart.robj, &adev->gart.ptr); in amdgpu_gart_table_vram_pin()
158 amdgpu_bo_unpin(adev->gart.robj); in amdgpu_gart_table_vram_pin()
159 amdgpu_bo_unreserve(adev->gart.robj); in amdgpu_gart_table_vram_pin()
160 adev->gart.table_addr = amdgpu_bo_gpu_offset(adev->gart.robj); in amdgpu_gart_table_vram_pin()
176 if (adev->gart.robj == NULL) { in amdgpu_gart_table_vram_unpin()
[all …]
Damdgpu_gem.c37 struct amdgpu_bo *robj = gem_to_amdgpu_bo(gobj); in amdgpu_gem_object_free() local
39 if (robj) { in amdgpu_gem_object_free()
40 amdgpu_mn_unregister(robj); in amdgpu_gem_object_free()
41 amdgpu_bo_unref(&robj); in amdgpu_gem_object_free()
377 struct amdgpu_bo *robj; in amdgpu_mode_dumb_mmap() local
383 robj = gem_to_amdgpu_bo(gobj); in amdgpu_mode_dumb_mmap()
384 if (amdgpu_ttm_tt_get_usermm(robj->tbo.ttm) || in amdgpu_mode_dumb_mmap()
385 (robj->flags & AMDGPU_GEM_CREATE_NO_CPU_ACCESS)) { in amdgpu_mode_dumb_mmap()
389 *offset_p = amdgpu_bo_mmap_offset(robj); in amdgpu_mode_dumb_mmap()
436 struct amdgpu_bo *robj; in amdgpu_gem_wait_idle_ioctl() local
[all …]
Damdgpu_bo_list.c53 amdgpu_bo_unref(&e->robj); in amdgpu_bo_list_free()
114 entry->robj = bo; in amdgpu_bo_list_create()
117 entry->tv.bo = &entry->robj->tbo; in amdgpu_bo_list_create()
118 entry->tv.shared = !entry->robj->prime_shared_count; in amdgpu_bo_list_create()
120 if (entry->robj->preferred_domains == AMDGPU_GEM_DOMAIN_GDS) in amdgpu_bo_list_create()
121 list->gds_obj = entry->robj; in amdgpu_bo_list_create()
122 if (entry->robj->preferred_domains == AMDGPU_GEM_DOMAIN_GWS) in amdgpu_bo_list_create()
123 list->gws_obj = entry->robj; in amdgpu_bo_list_create()
124 if (entry->robj->preferred_domains == AMDGPU_GEM_DOMAIN_OA) in amdgpu_bo_list_create()
125 list->oa_obj = entry->robj; in amdgpu_bo_list_create()
[all …]
Damdgpu_cs.c48 p->uf_entry.robj = amdgpu_bo_ref(gem_to_amdgpu_bo(gobj)); in amdgpu_cs_user_fence_chunk()
50 p->uf_entry.tv.bo = &p->uf_entry.robj->tbo; in amdgpu_cs_user_fence_chunk()
56 size = amdgpu_bo_size(p->uf_entry.robj); in amdgpu_cs_user_fence_chunk()
62 if (amdgpu_ttm_tt_get_usermm(p->uf_entry.robj->tbo.ttm)) { in amdgpu_cs_user_fence_chunk()
72 amdgpu_bo_unref(&p->uf_entry.robj); in amdgpu_cs_user_fence_chunk()
231 if (p->uf_entry.robj) in amdgpu_cs_parser_init()
460 struct amdgpu_bo *bo = candidate->robj; in amdgpu_cs_try_evict()
466 if (candidate->robj == validated) in amdgpu_cs_try_evict()
531 struct amdgpu_bo *bo = lobj->robj; in amdgpu_cs_list_validate()
606 if (p->uf_entry.robj && !p->uf_entry.robj->parent) in amdgpu_cs_parser_bos()
[all …]
Damdgpu_fb.c382 struct amdgpu_bo *robj; in amdgpu_fbdev_total_size() local
388 robj = gem_to_amdgpu_bo(adev->mode_info.rfbdev->rfb.base.obj[0]); in amdgpu_fbdev_total_size()
389 size += amdgpu_bo_size(robj); in amdgpu_fbdev_total_size()
393 bool amdgpu_fbdev_robj_is_fb(struct amdgpu_device *adev, struct amdgpu_bo *robj) in amdgpu_fbdev_robj_is_fb() argument
397 if (robj == gem_to_amdgpu_bo(adev->mode_info.rfbdev->rfb.base.obj[0])) in amdgpu_fbdev_robj_is_fb()
Damdgpu_gart.h44 struct amdgpu_bo *robj; member
Damdgpu_bo_list.h35 struct amdgpu_bo *robj; member
/Linux-v4.19/drivers/gpu/drm/nouveau/
Dnouveau_prime.c65 struct reservation_object *robj = attach->dmabuf->resv; in nouveau_gem_prime_import_sg_table() local
71 ww_mutex_lock(&robj->lock, NULL); in nouveau_gem_prime_import_sg_table()
73 sg, robj, &nvbo); in nouveau_gem_prime_import_sg_table()
74 ww_mutex_unlock(&robj->lock); in nouveau_gem_prime_import_sg_table()
Dnouveau_bo.h79 struct reservation_object *robj,
/Linux-v4.19/drivers/gpu/drm/etnaviv/
Detnaviv_gem.c462 struct reservation_object *robj = etnaviv_obj->resv; in etnaviv_gem_describe() local
473 fobj = rcu_dereference(robj->fence); in etnaviv_gem_describe()
483 fence = rcu_dereference(robj->fence_excl); in etnaviv_gem_describe()
570 struct reservation_object *robj, const struct etnaviv_gem_ops *ops, in etnaviv_gem_new_impl() argument
599 if (robj) { in etnaviv_gem_new_impl()
600 etnaviv_obj->resv = robj; in etnaviv_gem_new_impl()
661 struct reservation_object *robj, const struct etnaviv_gem_ops *ops, in etnaviv_gem_new_private() argument
667 ret = etnaviv_gem_new_impl(dev, size, flags, robj, ops, &obj); in etnaviv_gem_new_private()
Detnaviv_gem_submit.c179 struct reservation_object *robj = bo->obj->resv; in submit_fence_sync() local
182 ret = reservation_object_reserve_shared(robj); in submit_fence_sync()
191 ret = reservation_object_get_fences_rcu(robj, &bo->excl, in submit_fence_sync()
197 bo->excl = reservation_object_get_excl_rcu(robj); in submit_fence_sync()
Detnaviv_gem.h118 struct reservation_object *robj, const struct etnaviv_gem_ops *ops,
/Linux-v4.19/drivers/dma-buf/
Ddma-buf.c1016 struct reservation_object *robj; in dma_buf_debug_show() local
1047 robj = buf_obj->resv; in dma_buf_debug_show()
1049 seq = read_seqcount_begin(&robj->seq); in dma_buf_debug_show()
1051 fobj = rcu_dereference(robj->fence); in dma_buf_debug_show()
1053 fence = rcu_dereference(robj->fence_excl); in dma_buf_debug_show()
1054 if (!read_seqcount_retry(&robj->seq, seq)) in dma_buf_debug_show()

12