Home
last modified time | relevance | path

Searched full:engine (Results 1 – 25 of 1679) sorted by relevance

12345678910>>...68

/Linux-v5.4/crypto/
Dcrypto_engine.c3 * Handle async block request by crypto hardware engine.
12 #include <crypto/engine.h>
20 * @engine: the hardware engine
24 static void crypto_finalize_request(struct crypto_engine *engine, in crypto_finalize_request() argument
32 spin_lock_irqsave(&engine->queue_lock, flags); in crypto_finalize_request()
33 if (engine->cur_req == req) in crypto_finalize_request()
35 spin_unlock_irqrestore(&engine->queue_lock, flags); in crypto_finalize_request()
39 if (engine->cur_req_prepared && in crypto_finalize_request()
41 ret = enginectx->op.unprepare_request(engine, req); in crypto_finalize_request()
43 dev_err(engine->dev, "failed to unprepare request\n"); in crypto_finalize_request()
[all …]
/Linux-v5.4/drivers/gpu/drm/nouveau/nvkm/engine/disp/
DKbuild2 nvkm-y += nvkm/engine/disp/base.o
3 nvkm-y += nvkm/engine/disp/nv04.o
4 nvkm-y += nvkm/engine/disp/nv50.o
5 nvkm-y += nvkm/engine/disp/g84.o
6 nvkm-y += nvkm/engine/disp/g94.o
7 nvkm-y += nvkm/engine/disp/gt200.o
8 nvkm-y += nvkm/engine/disp/mcp77.o
9 nvkm-y += nvkm/engine/disp/gt215.o
10 nvkm-y += nvkm/engine/disp/mcp89.o
11 nvkm-y += nvkm/engine/disp/gf119.o
[all …]
/Linux-v5.4/drivers/gpu/drm/nouveau/nvkm/engine/gr/
DKbuild2 nvkm-y += nvkm/engine/gr/base.o
3 nvkm-y += nvkm/engine/gr/nv04.o
4 nvkm-y += nvkm/engine/gr/nv10.o
5 nvkm-y += nvkm/engine/gr/nv15.o
6 nvkm-y += nvkm/engine/gr/nv17.o
7 nvkm-y += nvkm/engine/gr/nv20.o
8 nvkm-y += nvkm/engine/gr/nv25.o
9 nvkm-y += nvkm/engine/gr/nv2a.o
10 nvkm-y += nvkm/engine/gr/nv30.o
11 nvkm-y += nvkm/engine/gr/nv34.o
[all …]
/Linux-v5.4/drivers/gpu/drm/nouveau/nvkm/core/
Dengine.c24 #include <core/engine.h>
31 nvkm_engine_chsw_load(struct nvkm_engine *engine) in nvkm_engine_chsw_load() argument
33 if (engine->func->chsw_load) in nvkm_engine_chsw_load()
34 return engine->func->chsw_load(engine); in nvkm_engine_chsw_load()
41 struct nvkm_engine *engine = *pengine; in nvkm_engine_unref() local
42 if (engine) { in nvkm_engine_unref()
43 mutex_lock(&engine->subdev.mutex); in nvkm_engine_unref()
44 if (--engine->usecount == 0) in nvkm_engine_unref()
45 nvkm_subdev_fini(&engine->subdev, false); in nvkm_engine_unref()
46 mutex_unlock(&engine->subdev.mutex); in nvkm_engine_unref()
[all …]
/Linux-v5.4/drivers/gpu/drm/i915/gt/
Dintel_engine_cs.c142 * intel_engine_context_size() - return the size of the context for an engine
144 * @class: engine class
146 * Each engine class may require a different amount of space for a context
149 * Return: size (in bytes) of an engine class specific context image
239 static void __sprint_engine_name(struct intel_engine_cs *engine) in __sprint_engine_name() argument
242 * Before we know what the uABI name for this engine will be, in __sprint_engine_name()
243 * we still would like to keep track of this engine in the debug logs. in __sprint_engine_name()
246 GEM_WARN_ON(snprintf(engine->name, sizeof(engine->name), "%s'%u", in __sprint_engine_name()
247 intel_engine_class_repr(engine->class), in __sprint_engine_name()
248 engine->instance) >= sizeof(engine->name)); in __sprint_engine_name()
[all …]
Dmock_engine.c46 static struct intel_ring *mock_ring(struct intel_engine_cs *engine) in mock_ring() argument
66 static struct i915_request *first_request(struct mock_engine *engine) in first_request() argument
68 return list_first_entry_or_null(&engine->hw_queue, in first_request()
79 intel_engine_queue_breadcrumbs(request->engine); in advance()
84 struct mock_engine *engine = from_timer(engine, t, hw_delay); in hw_delay_complete() local
88 spin_lock_irqsave(&engine->hw_lock, flags); in hw_delay_complete()
91 request = first_request(engine); in hw_delay_complete()
99 while ((request = first_request(engine))) { in hw_delay_complete()
101 mod_timer(&engine->hw_delay, in hw_delay_complete()
109 spin_unlock_irqrestore(&engine->hw_lock, flags); in hw_delay_complete()
[all …]
Dintel_engine_user.c38 void intel_engine_add_user(struct intel_engine_cs *engine) in intel_engine_add_user() argument
40 llist_add((struct llist_node *)&engine->uabi_node, in intel_engine_add_user()
41 (struct llist_head *)&engine->i915->uabi_engines); in intel_engine_add_user()
82 struct intel_engine_cs *engine = in sort_engines() local
83 container_of((struct rb_node *)pos, typeof(*engine), in sort_engines()
85 list_add((struct list_head *)&engine->uabi_node, engines); in sort_engines()
93 u8 engine; in set_scheduler_caps() member
102 struct intel_engine_cs *engine; in set_scheduler_caps() local
107 for_each_uabi_engine(engine, i915) { /* all engines must agree! */ in set_scheduler_caps()
110 if (engine->schedule) in set_scheduler_caps()
[all …]
Dintel_engine_pm.c17 struct intel_engine_cs *engine = in __engine_unpark() local
18 container_of(wf, typeof(*engine), wakeref); in __engine_unpark()
21 GEM_TRACE("%s\n", engine->name); in __engine_unpark()
23 intel_gt_pm_get(engine->gt); in __engine_unpark()
27 if (engine->default_state) in __engine_unpark()
28 map = i915_gem_object_pin_map(engine->default_state, in __engine_unpark()
31 engine->pinned_default_state = map; in __engine_unpark()
33 if (engine->unpark) in __engine_unpark()
34 engine->unpark(engine); in __engine_unpark()
36 intel_engine_init_hangcheck(engine); in __engine_unpark()
[all …]
Dintel_ringbuffer.c80 *cs++ = intel_gt_scratch_offset(rq->engine->gt, in gen2_render_ring_flush()
154 *cs++ = intel_gt_scratch_offset(rq->engine->gt, in gen4_render_ring_flush()
164 *cs++ = intel_gt_scratch_offset(rq->engine->gt, in gen4_render_ring_flush()
219 intel_gt_scratch_offset(rq->engine->gt, in gen6_emit_post_sync_nonzero_flush()
254 intel_gt_scratch_offset(rq->engine->gt, in gen6_render_ring_flush()
313 *cs++ = intel_gt_scratch_offset(rq->engine->gt, in gen6_rcs_emit_breadcrumb()
359 intel_gt_scratch_offset(rq->engine->gt, in gen7_render_ring_flush()
442 GEM_BUG_ON(rq->timeline->hwsp_ggtt != rq->engine->status_page.vma); in gen6_xcs_emit_breadcrumb()
462 GEM_BUG_ON(rq->timeline->hwsp_ggtt != rq->engine->status_page.vma); in gen7_xcs_emit_breadcrumb()
489 static void set_hwstam(struct intel_engine_cs *engine, u32 mask) in set_hwstam() argument
[all …]
Dintel_hangcheck.c53 static bool subunits_stuck(struct intel_engine_cs *engine) in subunits_stuck() argument
55 struct drm_i915_private *dev_priv = engine->i915; in subunits_stuck()
57 struct intel_instdone *accu_instdone = &engine->hangcheck.instdone; in subunits_stuck()
62 intel_engine_get_instdone(engine, &instdone); in subunits_stuck()
85 head_stuck(struct intel_engine_cs *engine, u64 acthd) in head_stuck() argument
87 if (acthd != engine->hangcheck.acthd) { in head_stuck()
90 memset(&engine->hangcheck.instdone, 0, in head_stuck()
91 sizeof(engine->hangcheck.instdone)); in head_stuck()
96 if (!subunits_stuck(engine)) in head_stuck()
103 engine_stuck(struct intel_engine_cs *engine, u64 acthd) in engine_stuck() argument
[all …]
Dintel_reset.c42 struct intel_engine_cs *engine = rq->engine; in engine_skip_context() local
48 lockdep_assert_held(&engine->active.lock); in engine_skip_context()
49 list_for_each_entry_continue(rq, &engine->active.requests, sched.link) in engine_skip_context()
120 rq->engine->name, in __i915_request_reset()
285 struct intel_engine_cs *engine; in gen6_reset_engines() local
301 for_each_engine_masked(engine, gt->i915, engine_mask, tmp) { in gen6_reset_engines()
302 GEM_BUG_ON(engine->id >= ARRAY_SIZE(hw_engine_mask)); in gen6_reset_engines()
303 hw_mask |= hw_engine_mask[engine->id]; in gen6_reset_engines()
310 static u32 gen11_lock_sfc(struct intel_engine_cs *engine) in gen11_lock_sfc() argument
312 struct intel_uncore *uncore = engine->uncore; in gen11_lock_sfc()
[all …]
Dintel_engine.h36 * ENGINE_READ(engine, REG_FOO);
41 * ENGINE_READ_IDX(engine, REG_BAR, i)
157 intel_read_status_page(const struct intel_engine_cs *engine, int reg) in intel_read_status_page() argument
160 return READ_ONCE(engine->status_page.addr[reg]); in intel_read_status_page()
164 intel_write_status_page(struct intel_engine_cs *engine, int reg, u32 value) in intel_write_status_page() argument
173 clflush(&engine->status_page.addr[reg]); in intel_write_status_page()
174 engine->status_page.addr[reg] = value; in intel_write_status_page()
175 clflush(&engine->status_page.addr[reg]); in intel_write_status_page()
178 WRITE_ONCE(engine->status_page.addr[reg], value); in intel_write_status_page()
210 intel_engine_create_ring(struct intel_engine_cs *engine, int size);
[all …]
Dintel_lrc.c49 * shouldn't we just need a set of those per engine command streamer? This is
51 * rings, the engine cs shifts to a new "ring buffer" with every context
66 * Now that ringbuffers belong per-context (and not per-engine, like before)
67 * and that contexts are uniquely tied to a given engine (and not reusable,
70 * - One ringbuffer per-engine inside each context.
71 * - One backing object per-engine inside each context.
75 * more complex, because we don't know at creation time which engine is going
80 * gets populated for a given engine once we receive an execbuffer. If later
82 * engine, we allocate/populate a new ringbuffer and context backing object and
99 * for the appropriate engine: this structure contains a copy of the context's
[all …]
Dselftest_workarounds.c33 } engine[I915_NUM_ENGINES]; member
39 struct intel_engine_cs *engine; in reference_lists_init() local
48 for_each_engine(engine, i915, id) { in reference_lists_init()
49 struct i915_wa_list *wal = &lists->engine[id].wa_list; in reference_lists_init()
51 wa_init_start(wal, "REF", engine->name); in reference_lists_init()
52 engine_init_workarounds(engine, wal); in reference_lists_init()
55 __intel_engine_init_ctx_wa(engine, in reference_lists_init()
56 &lists->engine[id].ctx_wa_list, in reference_lists_init()
64 struct intel_engine_cs *engine; in reference_lists_fini() local
67 for_each_engine(engine, i915, id) in reference_lists_fini()
[all …]
Dselftest_hangcheck.c131 hang_create_request(struct hang *h, struct intel_engine_cs *engine) in hang_create_request() argument
134 struct i915_address_space *vm = h->ctx->vm ?: &engine->gt->ggtt->vm; in hang_create_request()
175 rq = igt_request_alloc(h->ctx, engine); in hang_create_request()
244 intel_gt_chipset_flush(engine->gt); in hang_create_request()
246 if (rq->engine->emit_init_breadcrumb) { in hang_create_request()
247 err = rq->engine->emit_init_breadcrumb(rq); in hang_create_request()
256 err = rq->engine->emit_bb_start(rq, vma->node.start, PAGE_SIZE, flags); in hang_create_request()
305 struct intel_engine_cs *engine; in igt_hang_sanitycheck() local
317 for_each_engine(engine, gt->i915, id) { in igt_hang_sanitycheck()
321 if (!intel_engine_can_store_dword(engine)) in igt_hang_sanitycheck()
[all …]
Dselftest_context.c68 static int __live_context_size(struct intel_engine_cs *engine, in __live_context_size() argument
76 ce = intel_context_create(fixme, engine); in __live_context_size()
85 i915_coherent_map_type(engine->i915)); in __live_context_size()
104 if (HAS_EXECLISTS(engine->i915)) in __live_context_size()
107 vaddr += engine->context_size - I915_GTT_PAGE_SIZE; in __live_context_size()
122 rq = i915_request_create(engine->kernel_context); in __live_context_size()
132 pr_err("%s context overwrote trailing red-zone!", engine->name); in __live_context_size()
146 struct intel_engine_cs *engine; in live_context_size() local
164 for_each_engine(engine, gt->i915, id) { in live_context_size()
170 if (!engine->context_size) in live_context_size()
[all …]
/Linux-v5.4/drivers/gpu/drm/sun4i/
Dsunxi_engine.h25 * This callback allows to prepare our engine for an atomic
32 void (*atomic_begin)(struct sunxi_engine *engine,
49 int (*atomic_check)(struct sunxi_engine *engine,
61 void (*commit)(struct sunxi_engine *engine);
67 * the layers supported by that engine.
77 struct sunxi_engine *engine);
83 * engine. This is useful only for the composite output.
87 void (*apply_color_correction)(struct sunxi_engine *engine);
93 * engine. This is useful only for the composite output.
97 void (*disable_color_correction)(struct sunxi_engine *engine);
[all …]
/Linux-v5.4/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
DKbuild2 nvkm-y += nvkm/engine/fifo/base.o
3 nvkm-y += nvkm/engine/fifo/nv04.o
4 nvkm-y += nvkm/engine/fifo/nv10.o
5 nvkm-y += nvkm/engine/fifo/nv17.o
6 nvkm-y += nvkm/engine/fifo/nv40.o
7 nvkm-y += nvkm/engine/fifo/nv50.o
8 nvkm-y += nvkm/engine/fifo/g84.o
9 nvkm-y += nvkm/engine/fifo/gf100.o
10 nvkm-y += nvkm/engine/fifo/gk104.o
11 nvkm-y += nvkm/engine/fifo/gk110.o
[all …]
/Linux-v5.4/drivers/crypto/
Dpicoxcell_crypto.c81 struct spacc_engine *engine; member
95 struct spacc_engine *engine; member
138 struct spacc_engine *engine; member
146 struct spacc_engine *engine; member
186 static inline int spacc_fifo_cmd_full(struct spacc_engine *engine) in spacc_fifo_cmd_full() argument
188 u32 fifo_stat = readl(engine->regs + SPA_FIFO_STAT_REG_OFFSET); in spacc_fifo_cmd_full()
204 return is_cipher_ctx ? ctx->engine->cipher_ctx_base + in spacc_ctx_page_addr()
205 (indx * ctx->engine->cipher_pg_sz) : in spacc_ctx_page_addr()
206 ctx->engine->hash_key_base + (indx * ctx->engine->hash_pg_sz); in spacc_ctx_page_addr()
240 unsigned indx = ctx->engine->next_ctx++; in spacc_load_ctx()
[all …]
/Linux-v5.4/drivers/video/fbdev/via/
Daccel.c13 static int viafb_set_bpp(void __iomem *engine, u8 bpp) in viafb_set_bpp() argument
19 gemode = readl(engine + VIA_REG_GEMODE) & 0xfffffcfc; in viafb_set_bpp()
34 writel(gemode, engine + VIA_REG_GEMODE); in viafb_set_bpp()
39 static int hw_bitblt_1(void __iomem *engine, u8 op, u32 width, u32 height, in hw_bitblt_1() argument
79 ret = viafb_set_bpp(engine, dst_bpp); in hw_bitblt_1()
91 writel(tmp, engine + 0x08); in hw_bitblt_1()
100 writel(tmp, engine + 0x0C); in hw_bitblt_1()
108 writel(tmp, engine + 0x10); in hw_bitblt_1()
111 writel(fg_color, engine + 0x18); in hw_bitblt_1()
114 writel(bg_color, engine + 0x1C); in hw_bitblt_1()
[all …]
/Linux-v5.4/drivers/gpu/drm/nouveau/nvkm/engine/
DKbuild2 nvkm-y += nvkm/engine/falcon.o
3 nvkm-y += nvkm/engine/xtensa.o
5 include $(src)/nvkm/engine/bsp/Kbuild
6 include $(src)/nvkm/engine/ce/Kbuild
7 include $(src)/nvkm/engine/cipher/Kbuild
8 include $(src)/nvkm/engine/device/Kbuild
9 include $(src)/nvkm/engine/disp/Kbuild
10 include $(src)/nvkm/engine/dma/Kbuild
11 include $(src)/nvkm/engine/fifo/Kbuild
12 include $(src)/nvkm/engine/gr/Kbuild
[all …]
/Linux-v5.4/drivers/crypto/marvell/
Dcesa.c3 * Support for Marvell's Cryptographic Engine and Security Accelerator (CESA)
5 * driver supports the TDMA engine on platforms on which it is available.
38 mv_cesa_dequeue_req_locked(struct mv_cesa_engine *engine, in mv_cesa_dequeue_req_locked() argument
43 *backlog = crypto_get_backlog(&engine->queue); in mv_cesa_dequeue_req_locked()
44 req = crypto_dequeue_request(&engine->queue); in mv_cesa_dequeue_req_locked()
52 static void mv_cesa_rearm_engine(struct mv_cesa_engine *engine) in mv_cesa_rearm_engine() argument
58 spin_lock_bh(&engine->lock); in mv_cesa_rearm_engine()
59 if (!engine->req) { in mv_cesa_rearm_engine()
60 req = mv_cesa_dequeue_req_locked(engine, &backlog); in mv_cesa_rearm_engine()
61 engine->req = req; in mv_cesa_rearm_engine()
[all …]
/Linux-v5.4/include/crypto/
Dengine.h3 * Crypto engine API
22 * struct crypto_engine - crypto hardware engine
23 * @name: the engine name
24 * @idling: the engine is entering idle state
26 * @running: the engine is on working
28 * @list: link with the global crypto engine list
30 * @queue: the crypto queue of the engine
40 * @priv_data: the engine private data
57 int (*prepare_crypt_hardware)(struct crypto_engine *engine);
58 int (*unprepare_crypt_hardware)(struct crypto_engine *engine);
[all …]
/Linux-v5.4/drivers/dma/
DKconfig3 # DMA engine configuration
7 bool "DMA Engine support"
18 bool "DMA Engine debugging"
22 say N here. This enables DMA engine core and driver debugging.
25 bool "DMA Engine verbose debugging"
30 the DMA engine core and drivers.
61 tristate "Altera / Intel mSGDMA Engine"
73 provide DMA engine support. This includes the original ARM
113 tristate "Broadcom SBA RAID engine support"
122 Enable support for Broadcom SBA RAID Engine. The SBA RAID
[all …]
/Linux-v5.4/drivers/gpu/drm/nouveau/nvkm/engine/device/
Dpriv.h32 #include <engine/bsp.h>
33 #include <engine/ce.h>
34 #include <engine/cipher.h>
35 #include <engine/disp.h>
36 #include <engine/dma.h>
37 #include <engine/fifo.h>
38 #include <engine/gr.h>
39 #include <engine/mpeg.h>
40 #include <engine/mspdec.h>
41 #include <engine/msppp.h>
[all …]

12345678910>>...68