Home
last modified time | relevance | path

Searched refs:execlists (Results 1 – 16 of 16) sorted by relevance

/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/uc/
Dintel_guc_submission.c289 struct intel_engine_execlists * const execlists = &engine->execlists; in __guc_dequeue() local
290 struct i915_request **first = execlists->inflight; in __guc_dequeue()
291 struct i915_request ** const last_port = first + execlists->port_mask; in __guc_dequeue()
312 while ((rb = rb_first_cached(&execlists->queue))) { in __guc_dequeue()
323 port - execlists->inflight); in __guc_dequeue()
333 rb_erase_cached(&p->node, &execlists->queue); in __guc_dequeue()
337 execlists->queue_priority_hint = in __guc_dequeue()
340 *port = schedule_in(last, port - execlists->inflight); in __guc_dequeue()
344 execlists->active = execlists->inflight; in __guc_dequeue()
350 struct intel_engine_execlists * const execlists = &engine->execlists; in guc_submission_tasklet() local
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/gt/
Dintel_lrc.c437 static int queue_prio(const struct intel_engine_execlists *execlists) in queue_prio() argument
442 rb = rb_first_cached(&execlists->queue); in queue_prio()
485 if (engine->execlists.queue_priority_hint <= last_prio) in need_preempt()
525 return queue_prio(&engine->execlists) > last_prio; in need_preempt()
1137 GEM_BUG_ON(RB_EMPTY_ROOT(&engine->execlists.queue.rb_root)); in __unwind_incomplete_requests()
1162 execlists_unwind_incomplete_requests(struct intel_engine_execlists *execlists) in execlists_unwind_incomplete_requests() argument
1165 container_of(execlists, typeof(*engine), execlists); in execlists_unwind_incomplete_requests()
1367 ce->lrc.ccid |= engine->execlists.ccid; in __execlists_schedule_in()
1405 tasklet_hi_schedule(&ve->base.execlists.tasklet); in kick_siblings()
1526 static inline void write_desc(struct intel_engine_execlists *execlists, u64 desc, u32 port) in write_desc() argument
[all …]
Dintel_engine.h102 execlists_num_ports(const struct intel_engine_execlists * const execlists) in execlists_num_ports() argument
104 return execlists->port_mask + 1; in execlists_num_ports()
108 execlists_active(const struct intel_engine_execlists *execlists) in execlists_active() argument
112 cur = READ_ONCE(execlists->active); in execlists_active()
118 cur = READ_ONCE(execlists->active); in execlists_active()
127 execlists_active_lock_bh(struct intel_engine_execlists *execlists) in execlists_active_lock_bh() argument
130 tasklet_lock(&execlists->tasklet); in execlists_active_lock_bh()
134 execlists_active_unlock_bh(struct intel_engine_execlists *execlists) in execlists_active_unlock_bh() argument
136 tasklet_unlock(&execlists->tasklet); in execlists_active_unlock_bh()
141 execlists_unwind_incomplete_requests(struct intel_engine_execlists *execlists);
Dintel_engine_cs.c587 struct intel_engine_execlists * const execlists = &engine->execlists; in intel_engine_init_execlists() local
589 execlists->port_mask = 1; in intel_engine_init_execlists()
590 GEM_BUG_ON(!is_power_of_2(execlists_num_ports(execlists))); in intel_engine_init_execlists()
591 GEM_BUG_ON(execlists_num_ports(execlists) > EXECLIST_MAX_PORTS); in intel_engine_init_execlists()
593 memset(execlists->pending, 0, sizeof(execlists->pending)); in intel_engine_init_execlists()
594 execlists->active = in intel_engine_init_execlists()
595 memset(execlists->inflight, 0, sizeof(execlists->inflight)); in intel_engine_init_execlists()
597 execlists->queue_priority_hint = INT_MIN; in intel_engine_init_execlists()
598 execlists->queue = RB_ROOT_CACHED; in intel_engine_init_execlists()
931 tasklet_kill(&engine->execlists.tasklet); /* flush the callback */ in intel_engine_cleanup_common()
[all …]
Dintel_gt_irq.c39 WRITE_ONCE(engine->execlists.error_interrupt, eir); in cs_irq_handler()
45 WRITE_ONCE(engine->execlists.yield, in cs_irq_handler()
48 engine->execlists.yield); in cs_irq_handler()
49 if (del_timer(&engine->execlists.timer)) in cs_irq_handler()
62 tasklet_hi_schedule(&engine->execlists.tasklet); in cs_irq_handler()
Dsysfs_engines.c230 if (execlists_active(&engine->execlists)) in timeslice_store()
231 set_timer_ms(&engine->execlists.timer, duration); in timeslice_store()
334 if (READ_ONCE(engine->execlists.pending[0])) in preempt_timeout_store()
335 set_timer_ms(&engine->execlists.preempt, timeout); in preempt_timeout_store()
Dintel_engine_pm.c254 GEM_BUG_ON(engine->execlists.queue_priority_hint != INT_MIN); in __engine_park()
259 engine->execlists.no_priolist = false; in __engine_park()
Dselftest_lrc.c82 if (!READ_ONCE(engine->execlists.pending[0]) && is_active(rq)) in wait_for_submit()
102 if (READ_ONCE(engine->execlists.pending[0])) in wait_for_reset()
632 tasklet_disable(&engine->execlists.tasklet); in live_hold_reset()
634 engine->execlists.tasklet.func(engine->execlists.tasklet.data); in live_hold_reset()
635 GEM_BUG_ON(execlists_active(&engine->execlists) != rq); in live_hold_reset()
644 tasklet_enable(&engine->execlists.tasklet); in live_hold_reset()
1205 del_timer(&engine->execlists.timer); in live_timeslice_rewind()
1206 tasklet_hi_schedule(&engine->execlists.tasklet); in live_timeslice_rewind()
1373 GEM_BUG_ON(execlists_active(&engine->execlists) != rq); in live_timeslice_queue()
1384 } while (READ_ONCE(engine->execlists.pending[0])); in live_timeslice_queue()
[all …]
Dselftest_reset.c323 tasklet_disable(&engine->execlists.tasklet); in igt_atomic_engine_reset()
342 tasklet_enable(&engine->execlists.tasklet); in igt_atomic_engine_reset()
Dintel_engine_types.h475 struct intel_engine_execlists execlists; member
Dselftest_hangcheck.c1572 struct tasklet_struct * const t = &engine->execlists.tasklet; in __igt_atomic_reset_engine()
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/
Di915_scheduler.c43 static void assert_priolists(struct intel_engine_execlists * const execlists) in assert_priolists() argument
51 GEM_BUG_ON(rb_first_cached(&execlists->queue) != in assert_priolists()
52 rb_first(&execlists->queue.rb_root)); in assert_priolists()
55 for (rb = rb_first_cached(&execlists->queue); rb; rb = rb_next(rb)) { in assert_priolists()
74 struct intel_engine_execlists * const execlists = &engine->execlists; in i915_sched_lookup_priolist() local
81 assert_priolists(execlists); in i915_sched_lookup_priolist()
86 if (unlikely(execlists->no_priolist)) in i915_sched_lookup_priolist()
92 parent = &execlists->queue.rb_root.rb_node; in i915_sched_lookup_priolist()
107 p = &execlists->default_priolist; in i915_sched_lookup_priolist()
122 execlists->no_priolist = true; in i915_sched_lookup_priolist()
[all …]
DKconfig.profile46 when submitting a new context via execlists. If the current context
Di915_request.c356 return READ_ONCE(engine->execlists.active); in __engine_active()
Di915_gpu_error.c1251 const struct intel_engine_execlists * const el = &ee->engine->execlists; in engine_record_execlists()
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/selftests/
Di915_live_selftests.h46 selftest(execlists, intel_execlists_live_selftests)