| /kernel/linux/linux-5.10/crypto/ |
| D | crypto_engine.c | 3 * Handle async block request by crypto hardware engine. 13 #include <crypto/engine.h> 21 * @engine: the hardware engine 25 static void crypto_finalize_request(struct crypto_engine *engine, in crypto_finalize_request() argument 38 if (!engine->retry_support) { in crypto_finalize_request() 39 spin_lock_irqsave(&engine->queue_lock, flags); in crypto_finalize_request() 40 if (engine->cur_req == req) { in crypto_finalize_request() 42 engine->cur_req = NULL; in crypto_finalize_request() 44 spin_unlock_irqrestore(&engine->queue_lock, flags); in crypto_finalize_request() 47 if (finalize_req || engine->retry_support) { in crypto_finalize_request() [all …]
|
| /kernel/linux/linux-4.19/crypto/ |
| D | crypto_engine.c | 2 * Handle async block request by crypto hardware engine. 17 #include <crypto/engine.h> 25 * @engine: the hardware engine 29 static void crypto_finalize_request(struct crypto_engine *engine, in crypto_finalize_request() argument 37 spin_lock_irqsave(&engine->queue_lock, flags); in crypto_finalize_request() 38 if (engine->cur_req == req) in crypto_finalize_request() 40 spin_unlock_irqrestore(&engine->queue_lock, flags); in crypto_finalize_request() 44 if (engine->cur_req_prepared && in crypto_finalize_request() 46 ret = enginectx->op.unprepare_request(engine, req); in crypto_finalize_request() 48 dev_err(engine->dev, "failed to unprepare request\n"); in crypto_finalize_request() [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| D | Kbuild | 2 nvkm-y += nvkm/engine/disp/base.o 3 nvkm-y += nvkm/engine/disp/nv04.o 4 nvkm-y += nvkm/engine/disp/nv50.o 5 nvkm-y += nvkm/engine/disp/g84.o 6 nvkm-y += nvkm/engine/disp/g94.o 7 nvkm-y += nvkm/engine/disp/gt200.o 8 nvkm-y += nvkm/engine/disp/mcp77.o 9 nvkm-y += nvkm/engine/disp/gt215.o 10 nvkm-y += nvkm/engine/disp/mcp89.o 11 nvkm-y += nvkm/engine/disp/gf119.o [all …]
|
| /kernel/linux/linux-4.19/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| D | Kbuild | 1 nvkm-y += nvkm/engine/disp/base.o 2 nvkm-y += nvkm/engine/disp/nv04.o 3 nvkm-y += nvkm/engine/disp/nv50.o 4 nvkm-y += nvkm/engine/disp/g84.o 5 nvkm-y += nvkm/engine/disp/g94.o 6 nvkm-y += nvkm/engine/disp/gt200.o 7 nvkm-y += nvkm/engine/disp/mcp77.o 8 nvkm-y += nvkm/engine/disp/gt215.o 9 nvkm-y += nvkm/engine/disp/mcp89.o 10 nvkm-y += nvkm/engine/disp/gf119.o [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
| D | Kbuild | 2 nvkm-y += nvkm/engine/gr/base.o 3 nvkm-y += nvkm/engine/gr/nv04.o 4 nvkm-y += nvkm/engine/gr/nv10.o 5 nvkm-y += nvkm/engine/gr/nv15.o 6 nvkm-y += nvkm/engine/gr/nv17.o 7 nvkm-y += nvkm/engine/gr/nv20.o 8 nvkm-y += nvkm/engine/gr/nv25.o 9 nvkm-y += nvkm/engine/gr/nv2a.o 10 nvkm-y += nvkm/engine/gr/nv30.o 11 nvkm-y += nvkm/engine/gr/nv34.o [all …]
|
| /kernel/linux/linux-4.19/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
| D | Kbuild | 1 nvkm-y += nvkm/engine/gr/base.o 2 nvkm-y += nvkm/engine/gr/nv04.o 3 nvkm-y += nvkm/engine/gr/nv10.o 4 nvkm-y += nvkm/engine/gr/nv15.o 5 nvkm-y += nvkm/engine/gr/nv17.o 6 nvkm-y += nvkm/engine/gr/nv20.o 7 nvkm-y += nvkm/engine/gr/nv25.o 8 nvkm-y += nvkm/engine/gr/nv2a.o 9 nvkm-y += nvkm/engine/gr/nv30.o 10 nvkm-y += nvkm/engine/gr/nv34.o [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/ |
| D | intel_ring_submission.c | 48 static void set_hwstam(struct intel_engine_cs *engine, u32 mask) in set_hwstam() argument 54 if (engine->class == RENDER_CLASS) { in set_hwstam() 55 if (INTEL_GEN(engine->i915) >= 6) in set_hwstam() 61 intel_engine_set_hwsp_writemask(engine, mask); in set_hwstam() 64 static void set_hws_pga(struct intel_engine_cs *engine, phys_addr_t phys) in set_hws_pga() argument 69 if (INTEL_GEN(engine->i915) >= 4) in set_hws_pga() 72 intel_uncore_write(engine->uncore, HWS_PGA, addr); in set_hws_pga() 75 static struct page *status_page(struct intel_engine_cs *engine) in status_page() argument 77 struct drm_i915_gem_object *obj = engine->status_page.vma->obj; in status_page() 83 static void ring_setup_phys_status_page(struct intel_engine_cs *engine) in ring_setup_phys_status_page() argument [all …]
|
| D | intel_engine_heartbeat.c | 18 * While the engine is active, we send a periodic pulse along the engine 20 * is stuck, and we fail to preempt it, we declare the engine hung and 24 static bool next_heartbeat(struct intel_engine_cs *engine) in next_heartbeat() argument 28 delay = READ_ONCE(engine->props.heartbeat_interval_ms); in next_heartbeat() 35 mod_delayed_work(system_highpri_wq, &engine->heartbeat.work, delay); in next_heartbeat() 40 static void idle_pulse(struct intel_engine_cs *engine, struct i915_request *rq) in idle_pulse() argument 42 engine->wakeref_serial = READ_ONCE(engine->serial) + 1; in idle_pulse() 47 struct intel_engine_cs *engine) in show_heartbeat() argument 51 intel_engine_dump(engine, &p, in show_heartbeat() 53 engine->name, in show_heartbeat() [all …]
|
| D | intel_engine_cs.c | 144 * intel_engine_context_size() - return the size of the context for an engine 146 * @class: engine class 148 * Each engine class may require a different amount of space for a context 151 * Return: size (in bytes) of an engine class specific context image 242 static void __sprint_engine_name(struct intel_engine_cs *engine) in __sprint_engine_name() argument 245 * Before we know what the uABI name for this engine will be, in __sprint_engine_name() 246 * we still would like to keep track of this engine in the debug logs. in __sprint_engine_name() 249 GEM_WARN_ON(snprintf(engine->name, sizeof(engine->name), "%s'%u", in __sprint_engine_name() 250 intel_engine_class_repr(engine->class), in __sprint_engine_name() 251 engine->instance) >= sizeof(engine->name)); in __sprint_engine_name() [all …]
|
| D | mock_engine.c | 46 static struct intel_ring *mock_ring(struct intel_engine_cs *engine) in mock_ring() argument 84 static struct i915_request *first_request(struct mock_engine *engine) in first_request() argument 86 return list_first_entry_or_null(&engine->hw_queue, in first_request() 97 intel_engine_signal_breadcrumbs(request->engine); in advance() 102 struct mock_engine *engine = from_timer(engine, t, hw_delay); in hw_delay_complete() local 106 spin_lock_irqsave(&engine->hw_lock, flags); in hw_delay_complete() 109 request = first_request(engine); in hw_delay_complete() 117 while ((request = first_request(engine))) { in hw_delay_complete() 119 mod_timer(&engine->hw_delay, in hw_delay_complete() 127 spin_unlock_irqrestore(&engine->hw_lock, flags); in hw_delay_complete() [all …]
|
| D | selftest_engine_heartbeat.c | 32 static int engine_sync_barrier(struct intel_engine_cs *engine) in engine_sync_barrier() argument 34 return timeline_sync(engine->kernel_context->timeline); in engine_sync_barrier() 85 static int __live_idle_pulse(struct intel_engine_cs *engine, in __live_idle_pulse() argument 91 GEM_BUG_ON(!intel_engine_pm_is_awake(engine)); in __live_idle_pulse() 101 err = i915_active_acquire_preallocate_barrier(&p->active, engine); in __live_idle_pulse() 111 GEM_BUG_ON(llist_empty(&engine->barrier_tasks)); in __live_idle_pulse() 113 err = fn(engine); in __live_idle_pulse() 117 GEM_BUG_ON(!llist_empty(&engine->barrier_tasks)); in __live_idle_pulse() 119 if (engine_sync_barrier(engine)) { in __live_idle_pulse() 122 pr_err("%s: no heartbeat pulse?\n", engine->name); in __live_idle_pulse() [all …]
|
| D | intel_engine_user.c | 39 void intel_engine_add_user(struct intel_engine_cs *engine) in intel_engine_add_user() argument 41 llist_add((struct llist_node *)&engine->uabi_node, in intel_engine_add_user() 42 (struct llist_head *)&engine->i915->uabi_engines); in intel_engine_add_user() 84 struct intel_engine_cs *engine = in sort_engines() local 85 container_of((struct rb_node *)pos, typeof(*engine), in sort_engines() 87 list_add((struct list_head *)&engine->uabi_node, engines); in sort_engines() 95 u8 engine; in set_scheduler_caps() member 104 struct intel_engine_cs *engine; in set_scheduler_caps() local 109 for_each_uabi_engine(engine, i915) { /* all engines must agree! */ in set_scheduler_caps() 112 if (engine->schedule) in set_scheduler_caps() [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/core/ |
| D | engine.c | 24 #include <core/engine.h> 31 nvkm_engine_chsw_load(struct nvkm_engine *engine) in nvkm_engine_chsw_load() argument 33 if (engine->func->chsw_load) in nvkm_engine_chsw_load() 34 return engine->func->chsw_load(engine); in nvkm_engine_chsw_load() 41 struct nvkm_engine *engine = *pengine; in nvkm_engine_unref() local 42 if (engine) { in nvkm_engine_unref() 43 mutex_lock(&engine->subdev.mutex); in nvkm_engine_unref() 44 if (--engine->usecount == 0) in nvkm_engine_unref() 45 nvkm_subdev_fini(&engine->subdev, false); in nvkm_engine_unref() 46 mutex_unlock(&engine->subdev.mutex); in nvkm_engine_unref() [all …]
|
| /kernel/linux/linux-4.19/drivers/gpu/drm/nouveau/nvkm/core/ |
| D | engine.c | 24 #include <core/engine.h> 31 nvkm_engine_chsw_load(struct nvkm_engine *engine) in nvkm_engine_chsw_load() argument 33 if (engine->func->chsw_load) in nvkm_engine_chsw_load() 34 return engine->func->chsw_load(engine); in nvkm_engine_chsw_load() 41 struct nvkm_engine *engine = *pengine; in nvkm_engine_unref() local 42 if (engine) { in nvkm_engine_unref() 43 mutex_lock(&engine->subdev.mutex); in nvkm_engine_unref() 44 if (--engine->usecount == 0) in nvkm_engine_unref() 45 nvkm_subdev_fini(&engine->subdev, false); in nvkm_engine_unref() 46 mutex_unlock(&engine->subdev.mutex); in nvkm_engine_unref() [all …]
|
| /kernel/linux/linux-4.19/drivers/gpu/drm/i915/ |
| D | intel_engine_cs.c | 50 int (*init_legacy)(struct intel_engine_cs *engine); 51 int (*init_execlists)(struct intel_engine_cs *engine); 176 * ___intel_engine_context_size() - return the size of the context for an engine 178 * @class: engine class 180 * Each engine class may require a different amount of space for a context 183 * Return: size (in bytes) of an engine class specific context image 269 struct intel_engine_cs *engine; in intel_engine_setup() local 285 GEM_BUG_ON(dev_priv->engine[id]); in intel_engine_setup() 286 engine = kzalloc(sizeof(*engine), GFP_KERNEL); in intel_engine_setup() 287 if (!engine) in intel_engine_setup() [all …]
|
| D | intel_hangcheck.c | 28 ipehr_is_semaphore_wait(struct intel_engine_cs *engine, u32 ipehr) in ipehr_is_semaphore_wait() argument 36 semaphore_wait_to_signaller_ring(struct intel_engine_cs *engine, u32 ipehr, in semaphore_wait_to_signaller_ring() argument 39 struct drm_i915_private *dev_priv = engine->i915; in semaphore_wait_to_signaller_ring() 45 if (engine == signaller) in semaphore_wait_to_signaller_ring() 48 if (sync_bits == signaller->semaphore.mbox.wait[engine->hw_id]) in semaphore_wait_to_signaller_ring() 53 engine->name, ipehr); in semaphore_wait_to_signaller_ring() 59 semaphore_waits_for(struct intel_engine_cs *engine, u32 *seqno) in semaphore_waits_for() argument 61 struct drm_i915_private *dev_priv = engine->i915; in semaphore_waits_for() 84 if (engine->buffer == NULL) in semaphore_waits_for() 87 ipehr = I915_READ(RING_IPEHR(engine->mmio_base)); in semaphore_waits_for() [all …]
|
| D | intel_ringbuffer.c | 153 *cs++ = i915_ggtt_offset(rq->engine->scratch) | in gen4_render_ring_flush() 162 *cs++ = i915_ggtt_offset(rq->engine->scratch) | in gen4_render_ring_flush() 216 i915_ggtt_offset(rq->engine->scratch) + 2 * CACHELINE_BYTES; in intel_emit_post_sync_nonzero_flush() 250 i915_ggtt_offset(rq->engine->scratch) + 2 * CACHELINE_BYTES; in gen6_render_ring_flush() 320 i915_ggtt_offset(rq->engine->scratch) + 2 * CACHELINE_BYTES; in gen7_render_ring_flush() 378 static void ring_setup_phys_status_page(struct intel_engine_cs *engine) in ring_setup_phys_status_page() argument 380 struct drm_i915_private *dev_priv = engine->i915; in ring_setup_phys_status_page() 389 static void intel_ring_setup_status_page(struct intel_engine_cs *engine) in intel_ring_setup_status_page() argument 391 struct drm_i915_private *dev_priv = engine->i915; in intel_ring_setup_status_page() 398 switch (engine->id) { in intel_ring_setup_status_page() [all …]
|
| /kernel/linux/linux-4.19/drivers/gpu/drm/i915/selftests/ |
| D | mock_engine.c | 33 static struct mock_request *first_request(struct mock_engine *engine) in first_request() argument 35 return list_first_entry_or_null(&engine->hw_queue, in first_request() 40 static void advance(struct mock_engine *engine, in advance() argument 44 mock_seqno_advance(&engine->base, request->base.global_seqno); in advance() 49 struct mock_engine *engine = from_timer(engine, t, hw_delay); in hw_delay_complete() local 52 spin_lock(&engine->hw_lock); in hw_delay_complete() 55 request = first_request(engine); in hw_delay_complete() 57 advance(engine, request); in hw_delay_complete() 63 while ((request = first_request(engine))) { in hw_delay_complete() 65 mod_timer(&engine->hw_delay, jiffies + request->delay); in hw_delay_complete() [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/sun4i/ |
| D | sunxi_engine.h | 25 * This callback allows to prepare our engine for an atomic 32 void (*atomic_begin)(struct sunxi_engine *engine, 49 int (*atomic_check)(struct sunxi_engine *engine, 61 void (*commit)(struct sunxi_engine *engine); 67 * the layers supported by that engine. 77 struct sunxi_engine *engine); 83 * engine. This is useful only for the composite output. 87 void (*apply_color_correction)(struct sunxi_engine *engine); 93 * engine. This is useful only for the composite output. 97 void (*disable_color_correction)(struct sunxi_engine *engine); [all …]
|
| /kernel/linux/linux-4.19/drivers/gpu/drm/sun4i/ |
| D | sunxi_engine.h | 29 * This callback allows to prepare our engine for an atomic 36 void (*atomic_begin)(struct sunxi_engine *engine, 53 int (*atomic_check)(struct sunxi_engine *engine, 65 void (*commit)(struct sunxi_engine *engine); 71 * the layers supported by that engine. 81 struct sunxi_engine *engine); 87 * engine. This is useful only for the composite output. 91 void (*apply_color_correction)(struct sunxi_engine *engine); 97 * engine. This is useful only for the composite output. 101 void (*disable_color_correction)(struct sunxi_engine *engine); [all …]
|
| /kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
| D | Kbuild | 2 nvkm-y += nvkm/engine/fifo/base.o 3 nvkm-y += nvkm/engine/fifo/nv04.o 4 nvkm-y += nvkm/engine/fifo/nv10.o 5 nvkm-y += nvkm/engine/fifo/nv17.o 6 nvkm-y += nvkm/engine/fifo/nv40.o 7 nvkm-y += nvkm/engine/fifo/nv50.o 8 nvkm-y += nvkm/engine/fifo/g84.o 9 nvkm-y += nvkm/engine/fifo/gf100.o 10 nvkm-y += nvkm/engine/fifo/gk104.o 11 nvkm-y += nvkm/engine/fifo/gk110.o [all …]
|
| /kernel/linux/linux-4.19/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
| D | Kbuild | 1 nvkm-y += nvkm/engine/fifo/base.o 2 nvkm-y += nvkm/engine/fifo/nv04.o 3 nvkm-y += nvkm/engine/fifo/nv10.o 4 nvkm-y += nvkm/engine/fifo/nv17.o 5 nvkm-y += nvkm/engine/fifo/nv40.o 6 nvkm-y += nvkm/engine/fifo/nv50.o 7 nvkm-y += nvkm/engine/fifo/g84.o 8 nvkm-y += nvkm/engine/fifo/gf100.o 9 nvkm-y += nvkm/engine/fifo/gk104.o 10 nvkm-y += nvkm/engine/fifo/gk110.o [all …]
|
| /kernel/linux/linux-4.19/drivers/gpu/drm/amd/display/dc/i2caux/dce80/ |
| D | i2c_hw_engine_dce80.c | 32 #include "../engine.h" 88 * Cast pointer to 'struct engine *' 95 struct i2c_hw_engine_dce80 *engine) in disable_i2c_hw_engine() argument 97 const uint32_t addr = engine->addr.DC_I2C_DDCX_SETUP; in disable_i2c_hw_engine() 102 ctx = engine->base.base.base.ctx; in disable_i2c_hw_engine() 116 struct engine *engine) in release_engine() argument 118 struct i2c_hw_engine_dce80 *hw_engine = FROM_ENGINE(engine); in release_engine() 126 /* Restore original HW engine speed */ in release_engine() 132 value = dm_read_reg(engine->ctx, mmDC_I2C_ARBITRATION); in release_engine() 140 dm_write_reg(engine->ctx, mmDC_I2C_ARBITRATION, value); in release_engine() [all …]
|
| /kernel/linux/linux-4.19/drivers/crypto/ |
| D | picoxcell_crypto.c | 94 struct spacc_engine *engine; member 108 struct spacc_engine *engine; member 151 struct spacc_engine *engine; member 159 struct spacc_engine *engine; member 199 static inline int spacc_fifo_cmd_full(struct spacc_engine *engine) in spacc_fifo_cmd_full() argument 201 u32 fifo_stat = readl(engine->regs + SPA_FIFO_STAT_REG_OFFSET); in spacc_fifo_cmd_full() 217 return is_cipher_ctx ? ctx->engine->cipher_ctx_base + in spacc_ctx_page_addr() 218 (indx * ctx->engine->cipher_pg_sz) : in spacc_ctx_page_addr() 219 ctx->engine->hash_key_base + (indx * ctx->engine->hash_pg_sz); in spacc_ctx_page_addr() 253 unsigned indx = ctx->engine->next_ctx++; in spacc_load_ctx() [all …]
|
| /kernel/linux/linux-5.10/drivers/crypto/ |
| D | picoxcell_crypto.c | 81 struct spacc_engine *engine; member 96 struct spacc_engine *engine; member 139 struct spacc_engine *engine; member 147 struct spacc_engine *engine; member 187 static inline int spacc_fifo_cmd_full(struct spacc_engine *engine) in spacc_fifo_cmd_full() argument 189 u32 fifo_stat = readl(engine->regs + SPA_FIFO_STAT_REG_OFFSET); in spacc_fifo_cmd_full() 205 return is_cipher_ctx ? ctx->engine->cipher_ctx_base + in spacc_ctx_page_addr() 206 (indx * ctx->engine->cipher_pg_sz) : in spacc_ctx_page_addr() 207 ctx->engine->hash_key_base + (indx * ctx->engine->hash_pg_sz); in spacc_ctx_page_addr() 241 unsigned indx = ctx->engine->next_ctx++; in spacc_load_ctx() [all …]
|