Lines Matching refs:ce
24 struct intel_context *ce; in __engine_unpark() local
31 ce = engine->kernel_context; in __engine_unpark()
32 if (ce) { in __engine_unpark()
33 GEM_BUG_ON(test_bit(CONTEXT_VALID_BIT, &ce->flags)); in __engine_unpark()
36 if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_GEM) && ce->state) { in __engine_unpark()
37 struct drm_i915_gem_object *obj = ce->state->obj; in __engine_unpark()
49 ce->ops->reset(ce); in __engine_unpark()
61 static inline unsigned long __timeline_mark_lock(struct intel_context *ce) in __timeline_mark_lock() argument
66 mutex_acquire(&ce->timeline->mutex.dep_map, 2, 0, _THIS_IP_); in __timeline_mark_lock()
71 static inline void __timeline_mark_unlock(struct intel_context *ce, in __timeline_mark_unlock() argument
74 mutex_release(&ce->timeline->mutex.dep_map, _THIS_IP_); in __timeline_mark_unlock()
80 static inline unsigned long __timeline_mark_lock(struct intel_context *ce) in __timeline_mark_lock() argument
85 static inline void __timeline_mark_unlock(struct intel_context *ce, in __timeline_mark_unlock() argument
136 struct intel_context *ce = engine->kernel_context; in switch_to_kernel_context() local
145 GEM_BUG_ON(!intel_context_is_barrier(ce)); in switch_to_kernel_context()
146 GEM_BUG_ON(ce->timeline->hwsp_ggtt != engine->status_page.vma); in switch_to_kernel_context()
180 flags = __timeline_mark_lock(ce); in switch_to_kernel_context()
181 GEM_BUG_ON(atomic_read(&ce->timeline->active_count) < 0); in switch_to_kernel_context()
183 rq = __i915_request_create(ce, GFP_NOWAIT); in switch_to_kernel_context()
208 __queue_and_release_pm(rq, ce->timeline, engine); in switch_to_kernel_context()
212 __timeline_mark_unlock(ce, flags); in switch_to_kernel_context()