/Linux-v5.15/crypto/ |
D | crypto_engine.c | 25 static void crypto_finalize_request(struct crypto_engine *engine, in crypto_finalize_request() argument 38 if (!engine->retry_support) { in crypto_finalize_request() 39 spin_lock_irqsave(&engine->queue_lock, flags); in crypto_finalize_request() 40 if (engine->cur_req == req) { in crypto_finalize_request() 42 engine->cur_req = NULL; in crypto_finalize_request() 44 spin_unlock_irqrestore(&engine->queue_lock, flags); in crypto_finalize_request() 47 if (finalize_req || engine->retry_support) { in crypto_finalize_request() 51 ret = enginectx->op.unprepare_request(engine, req); in crypto_finalize_request() 53 dev_err(engine->dev, "failed to unprepare request\n"); in crypto_finalize_request() 58 kthread_queue_work(engine->kworker, &engine->pump_requests); in crypto_finalize_request() [all …]
|
/Linux-v5.15/drivers/gpu/drm/i915/gt/ |
D | intel_ring_submission.c | 26 static void set_hwstam(struct intel_engine_cs *engine, u32 mask) in set_hwstam() argument 32 if (engine->class == RENDER_CLASS) { in set_hwstam() 33 if (GRAPHICS_VER(engine->i915) >= 6) in set_hwstam() 39 intel_engine_set_hwsp_writemask(engine, mask); in set_hwstam() 42 static void set_hws_pga(struct intel_engine_cs *engine, phys_addr_t phys) in set_hws_pga() argument 47 if (GRAPHICS_VER(engine->i915) >= 4) in set_hws_pga() 50 intel_uncore_write(engine->uncore, HWS_PGA, addr); in set_hws_pga() 53 static struct page *status_page(struct intel_engine_cs *engine) in status_page() argument 55 struct drm_i915_gem_object *obj = engine->status_page.vma->obj; in status_page() 61 static void ring_setup_phys_status_page(struct intel_engine_cs *engine) in ring_setup_phys_status_page() argument [all …]
|
D | intel_engine_heartbeat.c | 23 static bool next_heartbeat(struct intel_engine_cs *engine) in next_heartbeat() argument 27 delay = READ_ONCE(engine->props.heartbeat_interval_ms); in next_heartbeat() 34 mod_delayed_work(system_highpri_wq, &engine->heartbeat.work, delay + 1); in next_heartbeat() 51 static void idle_pulse(struct intel_engine_cs *engine, struct i915_request *rq) in idle_pulse() argument 53 engine->wakeref_serial = READ_ONCE(engine->serial) + 1; in idle_pulse() 55 if (!engine->heartbeat.systole && intel_engine_has_heartbeat(engine)) in idle_pulse() 56 engine->heartbeat.systole = i915_request_get(rq); in idle_pulse() 62 idle_pulse(rq->engine, rq); in heartbeat_commit() 69 struct intel_engine_cs *engine) in show_heartbeat() argument 74 intel_engine_dump(engine, &p, in show_heartbeat() [all …]
|
D | intel_engine_cs.c | 255 static void __sprint_engine_name(struct intel_engine_cs *engine) in __sprint_engine_name() argument 262 GEM_WARN_ON(snprintf(engine->name, sizeof(engine->name), "%s'%u", in __sprint_engine_name() 263 intel_engine_class_repr(engine->class), in __sprint_engine_name() 264 engine->instance) >= sizeof(engine->name)); in __sprint_engine_name() 267 void intel_engine_set_hwsp_writemask(struct intel_engine_cs *engine, u32 mask) in intel_engine_set_hwsp_writemask() argument 273 if (GRAPHICS_VER(engine->i915) < 6 && engine->class != RENDER_CLASS) in intel_engine_set_hwsp_writemask() 276 if (GRAPHICS_VER(engine->i915) >= 3) in intel_engine_set_hwsp_writemask() 277 ENGINE_WRITE(engine, RING_HWSTAM, mask); in intel_engine_set_hwsp_writemask() 279 ENGINE_WRITE16(engine, RING_HWSTAM, mask); in intel_engine_set_hwsp_writemask() 282 static void intel_engine_sanitize_mmio(struct intel_engine_cs *engine) in intel_engine_sanitize_mmio() argument [all …]
|
D | mock_engine.c | 38 static struct intel_ring *mock_ring(struct intel_engine_cs *engine) in mock_ring() argument 76 static struct i915_request *first_request(struct mock_engine *engine) in first_request() argument 78 return list_first_entry_or_null(&engine->hw_queue, in first_request() 89 intel_engine_signal_breadcrumbs(request->engine); in advance() 94 struct mock_engine *engine = from_timer(engine, t, hw_delay); in hw_delay_complete() local 98 spin_lock_irqsave(&engine->hw_lock, flags); in hw_delay_complete() 101 request = first_request(engine); in hw_delay_complete() 109 while ((request = first_request(engine))) { in hw_delay_complete() 111 mod_timer(&engine->hw_delay, in hw_delay_complete() 119 spin_unlock_irqrestore(&engine->hw_lock, flags); in hw_delay_complete() [all …]
|
D | selftest_engine_heartbeat.c | 14 static void reset_heartbeat(struct intel_engine_cs *engine) in reset_heartbeat() argument 16 intel_engine_set_heartbeat(engine, in reset_heartbeat() 17 engine->defaults.heartbeat_interval_ms); in reset_heartbeat() 37 static int engine_sync_barrier(struct intel_engine_cs *engine) in engine_sync_barrier() argument 39 return timeline_sync(engine->kernel_context->timeline); in engine_sync_barrier() 90 static int __live_idle_pulse(struct intel_engine_cs *engine, in __live_idle_pulse() argument 96 GEM_BUG_ON(!intel_engine_pm_is_awake(engine)); in __live_idle_pulse() 106 err = i915_active_acquire_preallocate_barrier(&p->active, engine); in __live_idle_pulse() 116 GEM_BUG_ON(llist_empty(&engine->barrier_tasks)); in __live_idle_pulse() 118 err = fn(engine); in __live_idle_pulse() [all …]
|
D | intel_execlists_submission.c | 197 static struct virtual_engine *to_virtual_engine(struct intel_engine_cs *engine) in to_virtual_engine() argument 199 GEM_BUG_ON(!intel_engine_is_virtual(engine)); in to_virtual_engine() 200 return container_of(engine, struct virtual_engine, base); in to_virtual_engine() 233 static void ring_set_paused(const struct intel_engine_cs *engine, int state) in ring_set_paused() argument 241 engine->status_page.addr[I915_GEM_HWS_PREEMPT] = state; in ring_set_paused() 292 static bool need_preempt(const struct intel_engine_cs *engine, in need_preempt() argument 297 if (!intel_engine_has_semaphores(engine)) in need_preempt() 319 if (engine->sched_engine->queue_priority_hint <= last_prio) in need_preempt() 326 if (!list_is_last(&rq->sched.link, &engine->sched_engine->requests) && in need_preempt() 340 return max(virtual_prio(&engine->execlists), in need_preempt() [all …]
|
D | intel_engine.h | 132 intel_read_status_page(const struct intel_engine_cs *engine, int reg) in intel_read_status_page() argument 135 return READ_ONCE(engine->status_page.addr[reg]); in intel_read_status_page() 139 intel_write_status_page(struct intel_engine_cs *engine, int reg, u32 value) in intel_write_status_page() argument 148 clflush(&engine->status_page.addr[reg]); in intel_write_status_page() 149 engine->status_page.addr[reg] = value; in intel_write_status_page() 150 clflush(&engine->status_page.addr[reg]); in intel_write_status_page() 153 WRITE_ONCE(engine->status_page.addr[reg], value); in intel_write_status_page() 184 void intel_engine_stop(struct intel_engine_cs *engine); 185 void intel_engine_cleanup(struct intel_engine_cs *engine); 190 void intel_engine_free_request_pool(struct intel_engine_cs *engine); [all …]
|
D | intel_engine_user.c | 39 void intel_engine_add_user(struct intel_engine_cs *engine) in intel_engine_add_user() argument 41 llist_add((struct llist_node *)&engine->uabi_node, in intel_engine_add_user() 42 (struct llist_head *)&engine->i915->uabi_engines); in intel_engine_add_user() 84 struct intel_engine_cs *engine = in sort_engines() local 85 container_of((struct rb_node *)pos, typeof(*engine), in sort_engines() 87 list_add((struct list_head *)&engine->uabi_node, engines); in sort_engines() 95 u8 engine; in set_scheduler_caps() member 104 struct intel_engine_cs *engine; in set_scheduler_caps() local 109 for_each_uabi_engine(engine, i915) { /* all engines must agree! */ in set_scheduler_caps() 112 if (engine->sched_engine->schedule) in set_scheduler_caps() [all …]
|
D | intel_engine_pm.c | 26 int type = i915_coherent_map_type(ce->engine->i915, obj, true); in dbg_poison_ce() 44 struct intel_engine_cs *engine = in __engine_unpark() local 45 container_of(wf, typeof(*engine), wakeref); in __engine_unpark() 48 ENGINE_TRACE(engine, "\n"); in __engine_unpark() 50 intel_gt_pm_get(engine->gt); in __engine_unpark() 53 ce = engine->kernel_context; in __engine_unpark() 59 intel_engine_flush_submission(engine); in __engine_unpark() 75 if (engine->unpark) in __engine_unpark() 76 engine->unpark(engine); in __engine_unpark() 78 intel_breadcrumbs_unpark(engine->breadcrumbs); in __engine_unpark() [all …]
|
D | selftest_context.c | 75 static int __live_context_size(struct intel_engine_cs *engine) in __live_context_size() argument 82 ce = intel_context_create(engine); in __live_context_size() 91 i915_coherent_map_type(engine->i915, in __live_context_size() 111 vaddr += engine->context_size - I915_GTT_PAGE_SIZE; in __live_context_size() 126 rq = intel_engine_create_kernel_request(engine); in __live_context_size() 136 pr_err("%s context overwrote trailing red-zone!", engine->name); in __live_context_size() 150 struct intel_engine_cs *engine; in live_context_size() local 159 for_each_engine(engine, gt, id) { in live_context_size() 162 if (!engine->context_size) in live_context_size() 165 intel_engine_pm_get(engine); in live_context_size() [all …]
|
D | selftest_engine_pm.c | 74 struct intel_engine_cs *engine = ce->engine; in __measure_timestamps() local 75 u32 *sema = memset32(engine->status_page.addr + 1000, 0, 5); in __measure_timestamps() 76 u32 offset = i915_ggtt_offset(engine->status_page.vma); in __measure_timestamps() 94 cs = emit_srm(cs, RING_TIMESTAMP(engine->mmio_base), offset + 4000); in __measure_timestamps() 95 cs = emit_srm(cs, RING_CTX_TIMESTAMP(engine->mmio_base), offset + 4004); in __measure_timestamps() 100 cs = emit_srm(cs, RING_TIMESTAMP(engine->mmio_base), offset + 4016); in __measure_timestamps() 101 cs = emit_srm(cs, RING_CTX_TIMESTAMP(engine->mmio_base), offset + 4012); in __measure_timestamps() 106 intel_engine_flush_submission(engine); in __measure_timestamps() 130 engine->name, sema[1], sema[3], sema[0], sema[4]); in __measure_timestamps() 137 static int __live_engine_timestamps(struct intel_engine_cs *engine) in __live_engine_timestamps() argument [all …]
|
D | selftest_hangcheck.c | 117 hang_create_request(struct hang *h, struct intel_engine_cs *engine) in hang_create_request() argument 170 rq = igt_request_alloc(h->ctx, engine); in hang_create_request() 239 intel_gt_chipset_flush(engine->gt); in hang_create_request() 241 if (rq->engine->emit_init_breadcrumb) { in hang_create_request() 242 err = rq->engine->emit_init_breadcrumb(rq); in hang_create_request() 251 err = rq->engine->emit_bb_start(rq, vma->node.start, PAGE_SIZE, flags); in hang_create_request() 301 struct intel_engine_cs *engine; in igt_hang_sanitycheck() local 312 for_each_engine(engine, gt, id) { in igt_hang_sanitycheck() 316 if (!intel_engine_can_store_dword(engine)) in igt_hang_sanitycheck() 319 rq = hang_create_request(&h, engine); in igt_hang_sanitycheck() [all …]
|
D | intel_engine_pm.h | 14 intel_engine_pm_is_awake(const struct intel_engine_cs *engine) in intel_engine_pm_is_awake() argument 16 return intel_wakeref_is_active(&engine->wakeref); in intel_engine_pm_is_awake() 19 static inline void intel_engine_pm_get(struct intel_engine_cs *engine) in intel_engine_pm_get() argument 21 intel_wakeref_get(&engine->wakeref); in intel_engine_pm_get() 24 static inline bool intel_engine_pm_get_if_awake(struct intel_engine_cs *engine) in intel_engine_pm_get_if_awake() argument 26 return intel_wakeref_get_if_active(&engine->wakeref); in intel_engine_pm_get_if_awake() 29 static inline void intel_engine_pm_put(struct intel_engine_cs *engine) in intel_engine_pm_put() argument 31 intel_wakeref_put(&engine->wakeref); in intel_engine_pm_put() 34 static inline void intel_engine_pm_put_async(struct intel_engine_cs *engine) in intel_engine_pm_put_async() argument 36 intel_wakeref_put_async(&engine->wakeref); in intel_engine_pm_put_async() [all …]
|
D | intel_lrc.c | 21 const struct intel_engine_cs *engine, in set_offsets() argument 32 const u32 base = engine->mmio_base; in set_offsets() 50 if (GRAPHICS_VER(engine->i915) >= 11) in set_offsets() 73 if (GRAPHICS_VER(engine->i915) >= 11) in set_offsets() 534 static const u8 *reg_offsets(const struct intel_engine_cs *engine) in reg_offsets() argument 542 GEM_BUG_ON(GRAPHICS_VER(engine->i915) >= 12 && in reg_offsets() 543 !intel_engine_has_relative_mmio(engine)); in reg_offsets() 545 if (engine->class == RENDER_CLASS) { in reg_offsets() 546 if (GRAPHICS_VER_FULL(engine->i915) >= IP_VER(12, 50)) in reg_offsets() 548 else if (GRAPHICS_VER(engine->i915) >= 12) in reg_offsets() [all …]
|
D | intel_engine_stats.h | 16 static inline void intel_engine_context_in(struct intel_engine_cs *engine) in intel_engine_context_in() argument 20 if (engine->stats.active) { in intel_engine_context_in() 21 engine->stats.active++; in intel_engine_context_in() 27 write_seqcount_begin(&engine->stats.lock); in intel_engine_context_in() 29 engine->stats.start = ktime_get(); in intel_engine_context_in() 30 engine->stats.active++; in intel_engine_context_in() 32 write_seqcount_end(&engine->stats.lock); in intel_engine_context_in() 35 GEM_BUG_ON(!engine->stats.active); in intel_engine_context_in() 38 static inline void intel_engine_context_out(struct intel_engine_cs *engine) in intel_engine_context_out() argument 42 GEM_BUG_ON(!engine->stats.active); in intel_engine_context_out() [all …]
|
/Linux-v5.15/drivers/gpu/drm/nouveau/nvkm/core/ |
D | engine.c | 31 nvkm_engine_chsw_load(struct nvkm_engine *engine) in nvkm_engine_chsw_load() argument 33 if (engine->func->chsw_load) in nvkm_engine_chsw_load() 34 return engine->func->chsw_load(engine); in nvkm_engine_chsw_load() 41 struct nvkm_engine *engine = *pengine; in nvkm_engine_unref() local 42 if (engine) { in nvkm_engine_unref() 43 if (refcount_dec_and_mutex_lock(&engine->use.refcount, &engine->use.mutex)) { in nvkm_engine_unref() 44 nvkm_subdev_fini(&engine->subdev, false); in nvkm_engine_unref() 45 engine->use.enabled = false; in nvkm_engine_unref() 46 mutex_unlock(&engine->use.mutex); in nvkm_engine_unref() 53 nvkm_engine_ref(struct nvkm_engine *engine) in nvkm_engine_ref() argument [all …]
|
/Linux-v5.15/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
D | Kbuild | 2 nvkm-y += nvkm/engine/disp/base.o 3 nvkm-y += nvkm/engine/disp/nv04.o 4 nvkm-y += nvkm/engine/disp/nv50.o 5 nvkm-y += nvkm/engine/disp/g84.o 6 nvkm-y += nvkm/engine/disp/g94.o 7 nvkm-y += nvkm/engine/disp/gt200.o 8 nvkm-y += nvkm/engine/disp/mcp77.o 9 nvkm-y += nvkm/engine/disp/gt215.o 10 nvkm-y += nvkm/engine/disp/mcp89.o 11 nvkm-y += nvkm/engine/disp/gf119.o [all …]
|
/Linux-v5.15/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
D | Kbuild | 2 nvkm-y += nvkm/engine/gr/base.o 3 nvkm-y += nvkm/engine/gr/nv04.o 4 nvkm-y += nvkm/engine/gr/nv10.o 5 nvkm-y += nvkm/engine/gr/nv15.o 6 nvkm-y += nvkm/engine/gr/nv17.o 7 nvkm-y += nvkm/engine/gr/nv20.o 8 nvkm-y += nvkm/engine/gr/nv25.o 9 nvkm-y += nvkm/engine/gr/nv2a.o 10 nvkm-y += nvkm/engine/gr/nv30.o 11 nvkm-y += nvkm/engine/gr/nv34.o [all …]
|
/Linux-v5.15/drivers/gpu/drm/sun4i/ |
D | sunxi_engine.h | 32 void (*atomic_begin)(struct sunxi_engine *engine, 49 int (*atomic_check)(struct sunxi_engine *engine, 61 void (*commit)(struct sunxi_engine *engine); 77 struct sunxi_engine *engine); 87 void (*apply_color_correction)(struct sunxi_engine *engine); 97 void (*disable_color_correction)(struct sunxi_engine *engine); 110 void (*vblank_quirk)(struct sunxi_engine *engine); 137 sunxi_engine_commit(struct sunxi_engine *engine) in sunxi_engine_commit() argument 139 if (engine->ops && engine->ops->commit) in sunxi_engine_commit() 140 engine->ops->commit(engine); in sunxi_engine_commit() [all …]
|
/Linux-v5.15/drivers/video/fbdev/via/ |
D | accel.c | 13 static int viafb_set_bpp(void __iomem *engine, u8 bpp) in viafb_set_bpp() argument 19 gemode = readl(engine + VIA_REG_GEMODE) & 0xfffffcfc; in viafb_set_bpp() 34 writel(gemode, engine + VIA_REG_GEMODE); in viafb_set_bpp() 39 static int hw_bitblt_1(void __iomem *engine, u8 op, u32 width, u32 height, in hw_bitblt_1() argument 79 ret = viafb_set_bpp(engine, dst_bpp); in hw_bitblt_1() 91 writel(tmp, engine + 0x08); in hw_bitblt_1() 100 writel(tmp, engine + 0x0C); in hw_bitblt_1() 108 writel(tmp, engine + 0x10); in hw_bitblt_1() 111 writel(fg_color, engine + 0x18); in hw_bitblt_1() 114 writel(bg_color, engine + 0x1C); in hw_bitblt_1() [all …]
|
/Linux-v5.15/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
D | Kbuild | 2 nvkm-y += nvkm/engine/fifo/base.o 3 nvkm-y += nvkm/engine/fifo/nv04.o 4 nvkm-y += nvkm/engine/fifo/nv10.o 5 nvkm-y += nvkm/engine/fifo/nv17.o 6 nvkm-y += nvkm/engine/fifo/nv40.o 7 nvkm-y += nvkm/engine/fifo/nv50.o 8 nvkm-y += nvkm/engine/fifo/g84.o 9 nvkm-y += nvkm/engine/fifo/gf100.o 10 nvkm-y += nvkm/engine/fifo/gk104.o 11 nvkm-y += nvkm/engine/fifo/gk110.o [all …]
|
/Linux-v5.15/drivers/crypto/marvell/cesa/ |
D | cesa.c | 38 mv_cesa_dequeue_req_locked(struct mv_cesa_engine *engine, in mv_cesa_dequeue_req_locked() argument 43 *backlog = crypto_get_backlog(&engine->queue); in mv_cesa_dequeue_req_locked() 44 req = crypto_dequeue_request(&engine->queue); in mv_cesa_dequeue_req_locked() 52 static void mv_cesa_rearm_engine(struct mv_cesa_engine *engine) in mv_cesa_rearm_engine() argument 58 spin_lock_bh(&engine->lock); in mv_cesa_rearm_engine() 59 if (!engine->req) { in mv_cesa_rearm_engine() 60 req = mv_cesa_dequeue_req_locked(engine, &backlog); in mv_cesa_rearm_engine() 61 engine->req = req; in mv_cesa_rearm_engine() 63 spin_unlock_bh(&engine->lock); in mv_cesa_rearm_engine() 75 static int mv_cesa_std_process(struct mv_cesa_engine *engine, u32 status) in mv_cesa_std_process() argument [all …]
|
/Linux-v5.15/drivers/gpu/drm/nouveau/nvkm/engine/ |
D | Kbuild | 2 nvkm-y += nvkm/engine/falcon.o 3 nvkm-y += nvkm/engine/xtensa.o 5 include $(src)/nvkm/engine/bsp/Kbuild 6 include $(src)/nvkm/engine/ce/Kbuild 7 include $(src)/nvkm/engine/cipher/Kbuild 8 include $(src)/nvkm/engine/device/Kbuild 9 include $(src)/nvkm/engine/disp/Kbuild 10 include $(src)/nvkm/engine/dma/Kbuild 11 include $(src)/nvkm/engine/fifo/Kbuild 12 include $(src)/nvkm/engine/gr/Kbuild [all …]
|
/Linux-v5.15/drivers/gpu/drm/i915/selftests/ |
D | intel_scheduler_helpers.c | 17 int intel_selftest_modify_policy(struct intel_engine_cs *engine, in intel_selftest_modify_policy() argument 24 saved->reset = engine->i915->params.reset; in intel_selftest_modify_policy() 25 saved->flags = engine->flags; in intel_selftest_modify_policy() 26 saved->timeslice = engine->props.timeslice_duration_ms; in intel_selftest_modify_policy() 27 saved->preempt_timeout = engine->props.preempt_timeout_ms; in intel_selftest_modify_policy() 39 engine->i915->params.reset = 2; in intel_selftest_modify_policy() 40 engine->flags |= I915_ENGINE_WANT_FORCED_PREEMPTION; in intel_selftest_modify_policy() 41 engine->props.timeslice_duration_ms = REDUCED_TIMESLICE; in intel_selftest_modify_policy() 42 engine->props.preempt_timeout_ms = REDUCED_PREEMPT; in intel_selftest_modify_policy() 46 engine->props.preempt_timeout_ms = 0; in intel_selftest_modify_policy() [all …]
|