Lines Matching refs:obj

23 void i915_gem_object_free(struct drm_i915_gem_object *obj);
25 void i915_gem_object_init(struct drm_i915_gem_object *obj,
34 void __i915_gem_object_release_shmem(struct drm_i915_gem_object *obj,
38 int i915_gem_object_attach_phys(struct drm_i915_gem_object *obj, int align);
41 void i915_gem_free_object(struct drm_gem_object *obj);
46 __i915_gem_object_unset_pages(struct drm_i915_gem_object *obj);
47 void i915_gem_object_truncate(struct drm_i915_gem_object *obj);
72 struct drm_i915_gem_object *obj; in i915_gem_object_lookup() local
75 obj = i915_gem_object_lookup_rcu(file, handle); in i915_gem_object_lookup()
76 if (obj && !kref_get_unless_zero(&obj->base.refcount)) in i915_gem_object_lookup()
77 obj = NULL; in i915_gem_object_lookup()
80 return obj; in i915_gem_object_lookup()
89 i915_gem_object_get(struct drm_i915_gem_object *obj) in i915_gem_object_get() argument
91 drm_gem_object_get(&obj->base); in i915_gem_object_get()
92 return obj; in i915_gem_object_get()
97 i915_gem_object_put(struct drm_i915_gem_object *obj) in i915_gem_object_put() argument
99 __drm_gem_object_put(&obj->base); in i915_gem_object_put()
102 #define assert_object_held(obj) dma_resv_assert_held((obj)->base.resv) argument
104 static inline void i915_gem_object_lock(struct drm_i915_gem_object *obj) in i915_gem_object_lock() argument
106 dma_resv_lock(obj->base.resv, NULL); in i915_gem_object_lock()
110 i915_gem_object_lock_interruptible(struct drm_i915_gem_object *obj) in i915_gem_object_lock_interruptible() argument
112 return dma_resv_lock_interruptible(obj->base.resv, NULL); in i915_gem_object_lock_interruptible()
115 static inline void i915_gem_object_unlock(struct drm_i915_gem_object *obj) in i915_gem_object_unlock() argument
117 dma_resv_unlock(obj->base.resv); in i915_gem_object_unlock()
121 i915_gem_object_lock_fence(struct drm_i915_gem_object *obj);
122 void i915_gem_object_unlock_fence(struct drm_i915_gem_object *obj,
126 i915_gem_object_set_readonly(struct drm_i915_gem_object *obj) in i915_gem_object_set_readonly() argument
128 obj->base.vma_node.readonly = true; in i915_gem_object_set_readonly()
132 i915_gem_object_is_readonly(const struct drm_i915_gem_object *obj) in i915_gem_object_is_readonly() argument
134 return obj->base.vma_node.readonly; in i915_gem_object_is_readonly()
138 i915_gem_object_has_struct_page(const struct drm_i915_gem_object *obj) in i915_gem_object_has_struct_page() argument
140 return obj->ops->flags & I915_GEM_OBJECT_HAS_STRUCT_PAGE; in i915_gem_object_has_struct_page()
144 i915_gem_object_is_shrinkable(const struct drm_i915_gem_object *obj) in i915_gem_object_is_shrinkable() argument
146 return obj->ops->flags & I915_GEM_OBJECT_IS_SHRINKABLE; in i915_gem_object_is_shrinkable()
150 i915_gem_object_is_proxy(const struct drm_i915_gem_object *obj) in i915_gem_object_is_proxy() argument
152 return obj->ops->flags & I915_GEM_OBJECT_IS_PROXY; in i915_gem_object_is_proxy()
156 i915_gem_object_never_bind_ggtt(const struct drm_i915_gem_object *obj) in i915_gem_object_never_bind_ggtt() argument
158 return obj->ops->flags & I915_GEM_OBJECT_NO_GGTT; in i915_gem_object_never_bind_ggtt()
162 i915_gem_object_needs_async_cancel(const struct drm_i915_gem_object *obj) in i915_gem_object_needs_async_cancel() argument
164 return obj->ops->flags & I915_GEM_OBJECT_ASYNC_CANCEL; in i915_gem_object_needs_async_cancel()
168 i915_gem_object_is_framebuffer(const struct drm_i915_gem_object *obj) in i915_gem_object_is_framebuffer() argument
170 return READ_ONCE(obj->frontbuffer); in i915_gem_object_is_framebuffer()
174 i915_gem_object_get_tiling(const struct drm_i915_gem_object *obj) in i915_gem_object_get_tiling() argument
176 return obj->tiling_and_stride & TILING_MASK; in i915_gem_object_get_tiling()
180 i915_gem_object_is_tiled(const struct drm_i915_gem_object *obj) in i915_gem_object_is_tiled() argument
182 return i915_gem_object_get_tiling(obj) != I915_TILING_NONE; in i915_gem_object_is_tiled()
186 i915_gem_object_get_stride(const struct drm_i915_gem_object *obj) in i915_gem_object_get_stride() argument
188 return obj->tiling_and_stride & STRIDE_MASK; in i915_gem_object_get_stride()
199 i915_gem_object_get_tile_height(const struct drm_i915_gem_object *obj) in i915_gem_object_get_tile_height() argument
201 return i915_gem_tile_height(i915_gem_object_get_tiling(obj)); in i915_gem_object_get_tile_height()
205 i915_gem_object_get_tile_row_size(const struct drm_i915_gem_object *obj) in i915_gem_object_get_tile_row_size() argument
207 return (i915_gem_object_get_stride(obj) * in i915_gem_object_get_tile_row_size()
208 i915_gem_object_get_tile_height(obj)); in i915_gem_object_get_tile_row_size()
211 int i915_gem_object_set_tiling(struct drm_i915_gem_object *obj,
215 i915_gem_object_get_sg(struct drm_i915_gem_object *obj,
219 i915_gem_object_get_page(struct drm_i915_gem_object *obj,
223 i915_gem_object_get_dirty_page(struct drm_i915_gem_object *obj,
227 i915_gem_object_get_dma_address_len(struct drm_i915_gem_object *obj,
232 i915_gem_object_get_dma_address(struct drm_i915_gem_object *obj,
235 void __i915_gem_object_set_pages(struct drm_i915_gem_object *obj,
239 int ____i915_gem_object_get_pages(struct drm_i915_gem_object *obj);
240 int __i915_gem_object_get_pages(struct drm_i915_gem_object *obj);
243 i915_gem_object_pin_pages(struct drm_i915_gem_object *obj) in i915_gem_object_pin_pages() argument
245 might_lock(&obj->mm.lock); in i915_gem_object_pin_pages()
247 if (atomic_inc_not_zero(&obj->mm.pages_pin_count)) in i915_gem_object_pin_pages()
250 return __i915_gem_object_get_pages(obj); in i915_gem_object_pin_pages()
254 i915_gem_object_has_pages(struct drm_i915_gem_object *obj) in i915_gem_object_has_pages() argument
256 return !IS_ERR_OR_NULL(READ_ONCE(obj->mm.pages)); in i915_gem_object_has_pages()
260 __i915_gem_object_pin_pages(struct drm_i915_gem_object *obj) in __i915_gem_object_pin_pages() argument
262 GEM_BUG_ON(!i915_gem_object_has_pages(obj)); in __i915_gem_object_pin_pages()
264 atomic_inc(&obj->mm.pages_pin_count); in __i915_gem_object_pin_pages()
268 i915_gem_object_has_pinned_pages(struct drm_i915_gem_object *obj) in i915_gem_object_has_pinned_pages() argument
270 return atomic_read(&obj->mm.pages_pin_count); in i915_gem_object_has_pinned_pages()
274 __i915_gem_object_unpin_pages(struct drm_i915_gem_object *obj) in __i915_gem_object_unpin_pages() argument
276 GEM_BUG_ON(!i915_gem_object_has_pages(obj)); in __i915_gem_object_unpin_pages()
277 GEM_BUG_ON(!i915_gem_object_has_pinned_pages(obj)); in __i915_gem_object_unpin_pages()
279 atomic_dec(&obj->mm.pages_pin_count); in __i915_gem_object_unpin_pages()
283 i915_gem_object_unpin_pages(struct drm_i915_gem_object *obj) in i915_gem_object_unpin_pages() argument
285 __i915_gem_object_unpin_pages(obj); in i915_gem_object_unpin_pages()
293 int __i915_gem_object_put_pages(struct drm_i915_gem_object *obj,
295 void i915_gem_object_truncate(struct drm_i915_gem_object *obj);
296 void i915_gem_object_writeback(struct drm_i915_gem_object *obj);
322 void *__must_check i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
325 void __i915_gem_object_flush_map(struct drm_i915_gem_object *obj,
328 static inline void i915_gem_object_flush_map(struct drm_i915_gem_object *obj) in i915_gem_object_flush_map() argument
330 __i915_gem_object_flush_map(obj, 0, obj->base.size); in i915_gem_object_flush_map()
342 static inline void i915_gem_object_unpin_map(struct drm_i915_gem_object *obj) in i915_gem_object_unpin_map() argument
344 i915_gem_object_unpin_pages(obj); in i915_gem_object_unpin_map()
347 void __i915_gem_object_release_mmap(struct drm_i915_gem_object *obj);
348 void i915_gem_object_release_mmap(struct drm_i915_gem_object *obj);
351 i915_gem_object_flush_write_domain(struct drm_i915_gem_object *obj,
354 int i915_gem_object_prepare_read(struct drm_i915_gem_object *obj,
356 int i915_gem_object_prepare_write(struct drm_i915_gem_object *obj,
363 i915_gem_object_finish_access(struct drm_i915_gem_object *obj) in i915_gem_object_finish_access() argument
365 i915_gem_object_unpin_pages(obj); in i915_gem_object_finish_access()
366 i915_gem_object_unlock(obj); in i915_gem_object_finish_access()
370 i915_gem_object_last_write_engine(struct drm_i915_gem_object *obj) in i915_gem_object_last_write_engine() argument
376 fence = dma_resv_get_excl_rcu(obj->base.resv); in i915_gem_object_last_write_engine()
386 void i915_gem_object_set_cache_coherency(struct drm_i915_gem_object *obj,
388 void i915_gem_object_flush_if_display(struct drm_i915_gem_object *obj);
391 i915_gem_object_set_to_wc_domain(struct drm_i915_gem_object *obj, bool write);
393 i915_gem_object_set_to_gtt_domain(struct drm_i915_gem_object *obj, bool write);
395 i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, bool write);
397 i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj,
403 void i915_gem_object_make_unshrinkable(struct drm_i915_gem_object *obj);
404 void i915_gem_object_make_shrinkable(struct drm_i915_gem_object *obj);
405 void i915_gem_object_make_purgeable(struct drm_i915_gem_object *obj);
407 static inline bool cpu_write_needs_clflush(struct drm_i915_gem_object *obj) in cpu_write_needs_clflush() argument
409 if (obj->cache_dirty) in cpu_write_needs_clflush()
412 if (!(obj->cache_coherent & I915_BO_CACHE_COHERENT_FOR_WRITE)) in cpu_write_needs_clflush()
415 return obj->pin_global; /* currently in use by HW, keep flushed */ in cpu_write_needs_clflush()
418 static inline void __start_cpu_write(struct drm_i915_gem_object *obj) in __start_cpu_write() argument
420 obj->read_domains = I915_GEM_DOMAIN_CPU; in __start_cpu_write()
421 obj->write_domain = I915_GEM_DOMAIN_CPU; in __start_cpu_write()
422 if (cpu_write_needs_clflush(obj)) in __start_cpu_write()
423 obj->cache_dirty = true; in __start_cpu_write()
426 int i915_gem_object_wait(struct drm_i915_gem_object *obj,
429 int i915_gem_object_wait_priority(struct drm_i915_gem_object *obj,