/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/uc/ |
D | intel_guc_submission.c | 289 struct intel_engine_execlists * const execlists = &engine->execlists; in __guc_dequeue() local 290 struct i915_request **first = execlists->inflight; in __guc_dequeue() 291 struct i915_request ** const last_port = first + execlists->port_mask; in __guc_dequeue() 312 while ((rb = rb_first_cached(&execlists->queue))) { in __guc_dequeue() 323 port - execlists->inflight); in __guc_dequeue() 333 rb_erase_cached(&p->node, &execlists->queue); in __guc_dequeue() 337 execlists->queue_priority_hint = in __guc_dequeue() 340 *port = schedule_in(last, port - execlists->inflight); in __guc_dequeue() 344 execlists->active = execlists->inflight; in __guc_dequeue() 350 struct intel_engine_execlists * const execlists = &engine->execlists; in guc_submission_tasklet() local [all …]
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/ |
D | intel_lrc.c | 437 static int queue_prio(const struct intel_engine_execlists *execlists) in queue_prio() argument 442 rb = rb_first_cached(&execlists->queue); in queue_prio() 485 if (engine->execlists.queue_priority_hint <= last_prio) in need_preempt() 525 return queue_prio(&engine->execlists) > last_prio; in need_preempt() 1137 GEM_BUG_ON(RB_EMPTY_ROOT(&engine->execlists.queue.rb_root)); in __unwind_incomplete_requests() 1162 execlists_unwind_incomplete_requests(struct intel_engine_execlists *execlists) in execlists_unwind_incomplete_requests() argument 1165 container_of(execlists, typeof(*engine), execlists); in execlists_unwind_incomplete_requests() 1367 ce->lrc.ccid |= engine->execlists.ccid; in __execlists_schedule_in() 1405 tasklet_hi_schedule(&ve->base.execlists.tasklet); in kick_siblings() 1526 static inline void write_desc(struct intel_engine_execlists *execlists, u64 desc, u32 port) in write_desc() argument [all …]
|
D | intel_engine.h | 102 execlists_num_ports(const struct intel_engine_execlists * const execlists) in execlists_num_ports() argument 104 return execlists->port_mask + 1; in execlists_num_ports() 108 execlists_active(const struct intel_engine_execlists *execlists) in execlists_active() argument 112 cur = READ_ONCE(execlists->active); in execlists_active() 118 cur = READ_ONCE(execlists->active); in execlists_active() 127 execlists_active_lock_bh(struct intel_engine_execlists *execlists) in execlists_active_lock_bh() argument 130 tasklet_lock(&execlists->tasklet); in execlists_active_lock_bh() 134 execlists_active_unlock_bh(struct intel_engine_execlists *execlists) in execlists_active_unlock_bh() argument 136 tasklet_unlock(&execlists->tasklet); in execlists_active_unlock_bh() 141 execlists_unwind_incomplete_requests(struct intel_engine_execlists *execlists);
|
D | intel_engine_cs.c | 587 struct intel_engine_execlists * const execlists = &engine->execlists; in intel_engine_init_execlists() local 589 execlists->port_mask = 1; in intel_engine_init_execlists() 590 GEM_BUG_ON(!is_power_of_2(execlists_num_ports(execlists))); in intel_engine_init_execlists() 591 GEM_BUG_ON(execlists_num_ports(execlists) > EXECLIST_MAX_PORTS); in intel_engine_init_execlists() 593 memset(execlists->pending, 0, sizeof(execlists->pending)); in intel_engine_init_execlists() 594 execlists->active = in intel_engine_init_execlists() 595 memset(execlists->inflight, 0, sizeof(execlists->inflight)); in intel_engine_init_execlists() 597 execlists->queue_priority_hint = INT_MIN; in intel_engine_init_execlists() 598 execlists->queue = RB_ROOT_CACHED; in intel_engine_init_execlists() 931 tasklet_kill(&engine->execlists.tasklet); /* flush the callback */ in intel_engine_cleanup_common() [all …]
|
D | intel_gt_irq.c | 39 WRITE_ONCE(engine->execlists.error_interrupt, eir); in cs_irq_handler() 45 WRITE_ONCE(engine->execlists.yield, in cs_irq_handler() 48 engine->execlists.yield); in cs_irq_handler() 49 if (del_timer(&engine->execlists.timer)) in cs_irq_handler() 62 tasklet_hi_schedule(&engine->execlists.tasklet); in cs_irq_handler()
|
D | sysfs_engines.c | 230 if (execlists_active(&engine->execlists)) in timeslice_store() 231 set_timer_ms(&engine->execlists.timer, duration); in timeslice_store() 334 if (READ_ONCE(engine->execlists.pending[0])) in preempt_timeout_store() 335 set_timer_ms(&engine->execlists.preempt, timeout); in preempt_timeout_store()
|
D | intel_engine_pm.c | 254 GEM_BUG_ON(engine->execlists.queue_priority_hint != INT_MIN); in __engine_park() 259 engine->execlists.no_priolist = false; in __engine_park()
|
D | selftest_lrc.c | 82 if (!READ_ONCE(engine->execlists.pending[0]) && is_active(rq)) in wait_for_submit() 102 if (READ_ONCE(engine->execlists.pending[0])) in wait_for_reset() 632 tasklet_disable(&engine->execlists.tasklet); in live_hold_reset() 634 engine->execlists.tasklet.func(engine->execlists.tasklet.data); in live_hold_reset() 635 GEM_BUG_ON(execlists_active(&engine->execlists) != rq); in live_hold_reset() 644 tasklet_enable(&engine->execlists.tasklet); in live_hold_reset() 1205 del_timer(&engine->execlists.timer); in live_timeslice_rewind() 1206 tasklet_hi_schedule(&engine->execlists.tasklet); in live_timeslice_rewind() 1373 GEM_BUG_ON(execlists_active(&engine->execlists) != rq); in live_timeslice_queue() 1384 } while (READ_ONCE(engine->execlists.pending[0])); in live_timeslice_queue() [all …]
|
D | selftest_reset.c | 323 tasklet_disable(&engine->execlists.tasklet); in igt_atomic_engine_reset() 342 tasklet_enable(&engine->execlists.tasklet); in igt_atomic_engine_reset()
|
D | intel_engine_types.h | 475 struct intel_engine_execlists execlists; member
|
D | selftest_hangcheck.c | 1572 struct tasklet_struct * const t = &engine->execlists.tasklet; in __igt_atomic_reset_engine()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/ |
D | i915_scheduler.c | 43 static void assert_priolists(struct intel_engine_execlists * const execlists) in assert_priolists() argument 51 GEM_BUG_ON(rb_first_cached(&execlists->queue) != in assert_priolists() 52 rb_first(&execlists->queue.rb_root)); in assert_priolists() 55 for (rb = rb_first_cached(&execlists->queue); rb; rb = rb_next(rb)) { in assert_priolists() 74 struct intel_engine_execlists * const execlists = &engine->execlists; in i915_sched_lookup_priolist() local 81 assert_priolists(execlists); in i915_sched_lookup_priolist() 86 if (unlikely(execlists->no_priolist)) in i915_sched_lookup_priolist() 92 parent = &execlists->queue.rb_root.rb_node; in i915_sched_lookup_priolist() 107 p = &execlists->default_priolist; in i915_sched_lookup_priolist() 122 execlists->no_priolist = true; in i915_sched_lookup_priolist() [all …]
|
D | Kconfig.profile | 46 when submitting a new context via execlists. If the current context
|
D | i915_request.c | 356 return READ_ONCE(engine->execlists.active); in __engine_active()
|
D | i915_gpu_error.c | 1251 const struct intel_engine_execlists * const el = &ee->engine->execlists; in engine_record_execlists()
|
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/ |
D | i915_live_selftests.h | 46 selftest(execlists, intel_execlists_live_selftests)
|