/* * Copyright © 2016 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #include "config.h" #include #include #include #include #include "igt.h" #include "igt_gpu_power.h" #include "igt_rand.h" #include "igt_sysfs.h" #include "igt_vgem.h" #include "i915/gem_ring.h" #define LO 0 #define HI 1 #define NOISE 2 #define MAX_PRIO LOCAL_I915_CONTEXT_MAX_USER_PRIORITY #define MIN_PRIO LOCAL_I915_CONTEXT_MIN_USER_PRIORITY #define MAX_ELSP_QLEN 16 #define MAX_ENGINES 16 #define MAX_CONTEXTS 1024 #define LOCAL_I915_EXEC_BSD_SHIFT (13) #define LOCAL_I915_EXEC_BSD_MASK (3 << LOCAL_I915_EXEC_BSD_SHIFT) #define ENGINE_MASK (I915_EXEC_RING_MASK | LOCAL_I915_EXEC_BSD_MASK) #define MI_SEMAPHORE_WAIT (0x1c << 23) #define MI_SEMAPHORE_POLL (1 << 15) #define MI_SEMAPHORE_SAD_GT_SDD (0 << 12) #define MI_SEMAPHORE_SAD_GTE_SDD (1 << 12) #define MI_SEMAPHORE_SAD_LT_SDD (2 << 12) #define MI_SEMAPHORE_SAD_LTE_SDD (3 << 12) #define MI_SEMAPHORE_SAD_EQ_SDD (4 << 12) #define MI_SEMAPHORE_SAD_NEQ_SDD (5 << 12) IGT_TEST_DESCRIPTION("Check that we can control the order of execution"); static inline uint32_t __sync_read_u32(int fd, uint32_t handle, uint64_t offset) { uint32_t value; gem_set_domain(fd, handle, /* No write hazard lies! */ I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT); gem_read(fd, handle, offset, &value, sizeof(value)); return value; } static inline void __sync_read_u32_count(int fd, uint32_t handle, uint32_t *dst, uint64_t size) { gem_set_domain(fd, handle, /* No write hazard lies! */ I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT); gem_read(fd, handle, 0, dst, size); } static uint32_t __store_dword(int fd, uint32_t ctx, unsigned ring, uint32_t target, uint32_t offset, uint32_t value, uint32_t cork, unsigned write_domain) { const int gen = intel_gen(intel_get_drm_devid(fd)); struct drm_i915_gem_exec_object2 obj[3]; struct drm_i915_gem_relocation_entry reloc; struct drm_i915_gem_execbuffer2 execbuf; uint32_t batch[16]; int i; memset(&execbuf, 0, sizeof(execbuf)); execbuf.buffers_ptr = to_user_pointer(obj + !cork); execbuf.buffer_count = 2 + !!cork; execbuf.flags = ring; if (gen < 6) execbuf.flags |= I915_EXEC_SECURE; execbuf.rsvd1 = ctx; memset(obj, 0, sizeof(obj)); obj[0].handle = cork; obj[1].handle = target; obj[2].handle = gem_create(fd, 4096); memset(&reloc, 0, sizeof(reloc)); reloc.target_handle = obj[1].handle; reloc.presumed_offset = 0; reloc.offset = sizeof(uint32_t); reloc.delta = offset; reloc.read_domains = I915_GEM_DOMAIN_INSTRUCTION; reloc.write_domain = write_domain; obj[2].relocs_ptr = to_user_pointer(&reloc); obj[2].relocation_count = 1; i = 0; batch[i] = MI_STORE_DWORD_IMM | (gen < 6 ? 1 << 22 : 0); if (gen >= 8) { batch[++i] = offset; batch[++i] = 0; } else if (gen >= 4) { batch[++i] = 0; batch[++i] = offset; reloc.offset += sizeof(uint32_t); } else { batch[i]--; batch[++i] = offset; } batch[++i] = value; batch[++i] = MI_BATCH_BUFFER_END; gem_write(fd, obj[2].handle, 0, batch, sizeof(batch)); gem_execbuf(fd, &execbuf); return obj[2].handle; } static void store_dword(int fd, uint32_t ctx, unsigned ring, uint32_t target, uint32_t offset, uint32_t value, uint32_t cork, unsigned write_domain) { gem_close(fd, __store_dword(fd, ctx, ring, target, offset, value, cork, write_domain)); } static uint32_t create_highest_priority(int fd) { uint32_t ctx = gem_context_create(fd); /* * If there is no priority support, all contexts will have equal * priority (and therefore the max user priority), so no context * can overtake us, and we effectively can form a plug. */ __gem_context_set_priority(fd, ctx, MAX_PRIO); return ctx; } static void unplug_show_queue(int fd, struct igt_cork *c, unsigned int engine) { igt_spin_t *spin[MAX_ELSP_QLEN]; int max = MAX_ELSP_QLEN; /* If no scheduler, all batches are emitted in submission order */ if (!gem_scheduler_enabled(fd)) max = 1; for (int n = 0; n < max; n++) { const struct igt_spin_factory opts = { .ctx = create_highest_priority(fd), .engine = engine, }; spin[n] = __igt_spin_factory(fd, &opts); gem_context_destroy(fd, opts.ctx); } igt_cork_unplug(c); /* batches will now be queued on the engine */ igt_debugfs_dump(fd, "i915_engine_info"); for (int n = 0; n < max; n++) igt_spin_free(fd, spin[n]); } static void fifo(int fd, unsigned ring) { IGT_CORK_HANDLE(cork); uint32_t scratch, plug; uint32_t result; scratch = gem_create(fd, 4096); plug = igt_cork_plug(&cork, fd); /* Same priority, same timeline, final result will be the second eb */ store_dword(fd, 0, ring, scratch, 0, 1, plug, 0); store_dword(fd, 0, ring, scratch, 0, 2, plug, 0); unplug_show_queue(fd, &cork, ring); gem_close(fd, plug); result = __sync_read_u32(fd, scratch, 0); gem_close(fd, scratch); igt_assert_eq_u32(result, 2); } static void independent(int fd, unsigned int engine) { IGT_CORK_HANDLE(cork); uint32_t scratch, plug, batch; igt_spin_t *spin = NULL; unsigned int other; uint32_t *ptr; igt_require(engine != 0); scratch = gem_create(fd, 4096); ptr = gem_mmap__gtt(fd, scratch, 4096, PROT_READ); igt_assert_eq(ptr[0], 0); plug = igt_cork_plug(&cork, fd); /* Check that we can submit to engine while all others are blocked */ for_each_physical_engine(fd, other) { if (other == engine) continue; if (!gem_can_store_dword(fd, other)) continue; if (spin == NULL) { spin = __igt_spin_new(fd, .engine = other); } else { struct drm_i915_gem_execbuffer2 eb = { .buffer_count = 1, .buffers_ptr = to_user_pointer(&spin->obj[IGT_SPIN_BATCH]), .flags = other, }; gem_execbuf(fd, &eb); } store_dword(fd, 0, other, scratch, 0, other, plug, 0); } igt_require(spin); /* Same priority, but different timeline (as different engine) */ batch = __store_dword(fd, 0, engine, scratch, 0, engine, plug, 0); unplug_show_queue(fd, &cork, engine); gem_close(fd, plug); gem_sync(fd, batch); igt_assert(!gem_bo_busy(fd, batch)); igt_assert(gem_bo_busy(fd, spin->handle)); gem_close(fd, batch); /* Only the local engine should be free to complete. */ igt_assert(gem_bo_busy(fd, scratch)); igt_assert_eq(ptr[0], engine); igt_spin_free(fd, spin); gem_quiescent_gpu(fd); /* And we expect the others to have overwritten us, order unspecified */ igt_assert(!gem_bo_busy(fd, scratch)); igt_assert_neq(ptr[0], engine); munmap(ptr, 4096); gem_close(fd, scratch); } static void smoketest(int fd, unsigned ring, unsigned timeout) { const int ncpus = sysconf(_SC_NPROCESSORS_ONLN); unsigned engines[MAX_ENGINES]; unsigned nengine; unsigned engine; uint32_t scratch; uint32_t result[2 * ncpus]; nengine = 0; if (ring == ALL_ENGINES) { for_each_physical_engine(fd, engine) if (gem_can_store_dword(fd, engine)) engines[nengine++] = engine; } else { if (gem_can_store_dword(fd, ring)) engines[nengine++] = ring; } igt_require(nengine); scratch = gem_create(fd, 4096); igt_fork(child, ncpus) { unsigned long count = 0; uint32_t ctx; hars_petruska_f54_1_random_perturb(child); ctx = gem_context_create(fd); igt_until_timeout(timeout) { int prio; prio = hars_petruska_f54_1_random_unsafe_max(MAX_PRIO - MIN_PRIO) + MIN_PRIO; gem_context_set_priority(fd, ctx, prio); engine = engines[hars_petruska_f54_1_random_unsafe_max(nengine)]; store_dword(fd, ctx, engine, scratch, 8*child + 0, ~child, 0, 0); for (unsigned int step = 0; step < 8; step++) store_dword(fd, ctx, engine, scratch, 8*child + 4, count++, 0, 0); } gem_context_destroy(fd, ctx); } igt_waitchildren(); __sync_read_u32_count(fd, scratch, result, sizeof(result)); gem_close(fd, scratch); for (unsigned n = 0; n < ncpus; n++) { igt_assert_eq_u32(result[2 * n], ~n); /* * Note this count is approximate due to unconstrained * ordering of the dword writes between engines. * * Take the result with a pinch of salt. */ igt_info("Child[%d] completed %u cycles\n", n, result[(2 * n) + 1]); } } static uint32_t __batch_create(int i915, uint32_t offset) { const uint32_t bbe = MI_BATCH_BUFFER_END; uint32_t handle; handle = gem_create(i915, ALIGN(offset + 4, 4096)); gem_write(i915, handle, offset, &bbe, sizeof(bbe)); return handle; } static uint32_t batch_create(int i915) { return __batch_create(i915, 0); } static void semaphore_userlock(int i915) { struct drm_i915_gem_exec_object2 obj = { .handle = batch_create(i915), }; igt_spin_t *spin = NULL; unsigned int engine; uint32_t scratch; igt_require(gem_scheduler_has_semaphores(i915)); /* * Given the use of semaphores to govern parallel submission * of nearly-ready work to HW, we still want to run actually * ready work immediately. Without semaphores, the dependent * work wouldn't be submitted so our ready work will run. */ scratch = gem_create(i915, 4096); for_each_physical_engine(i915, engine) { if (!spin) { spin = igt_spin_new(i915, .dependency = scratch, .engine = engine); } else { uint64_t saved = spin->execbuf.flags; spin->execbuf.flags &= ~ENGINE_MASK; spin->execbuf.flags |= engine; gem_execbuf(i915, &spin->execbuf); spin->execbuf.flags = saved; } } igt_require(spin); gem_close(i915, scratch); /* * On all dependent engines, the request may be executing (busywaiting * on a HW semaphore) but it should not prevent any real work from * taking precedence. */ scratch = gem_context_create(i915); for_each_physical_engine(i915, engine) { struct drm_i915_gem_execbuffer2 execbuf = { .buffers_ptr = to_user_pointer(&obj), .buffer_count = 1, .flags = engine, .rsvd1 = scratch, }; if (engine == (spin->execbuf.flags & ENGINE_MASK)) continue; gem_execbuf(i915, &execbuf); } gem_context_destroy(i915, scratch); gem_sync(i915, obj.handle); /* to hang unless we can preempt */ gem_close(i915, obj.handle); igt_spin_free(i915, spin); } static void semaphore_codependency(int i915) { struct { igt_spin_t *xcs, *rcs; } task[2]; unsigned int engine; int i; /* * Consider two tasks, task A runs on (xcs0, rcs0) and task B * on (xcs1, rcs0). That is they must both run a dependent * batch on rcs0, after first running in parallel on separate * engines. To maximise throughput, we want the shorter xcs task * to start on rcs first. However, if we insert semaphores we may * pick wrongly and end up running the requests in the least * optimal order. */ i = 0; for_each_physical_engine(i915, engine) { uint32_t ctx; if (engine == I915_EXEC_RENDER) continue; if (!gem_can_store_dword(i915, engine)) continue; ctx = gem_context_create(i915); task[i].xcs = __igt_spin_new(i915, .ctx = ctx, .engine = engine, .flags = IGT_SPIN_POLL_RUN); igt_spin_busywait_until_started(task[i].xcs); /* Common rcs tasks will be queued in FIFO */ task[i].rcs = __igt_spin_new(i915, .ctx = ctx, .engine = I915_EXEC_RENDER, .dependency = task[i].xcs->handle); gem_context_destroy(i915, ctx); if (++i == ARRAY_SIZE(task)) break; } igt_require(i == ARRAY_SIZE(task)); /* Since task[0] was queued first, it will be first in queue for rcs */ igt_spin_end(task[1].xcs); igt_spin_end(task[1].rcs); gem_sync(i915, task[1].rcs->handle); /* to hang if task[0] hogs rcs */ for (i = 0; i < ARRAY_SIZE(task); i++) { igt_spin_free(i915, task[i].xcs); igt_spin_free(i915, task[i].rcs); } } static unsigned int offset_in_page(void *addr) { return (uintptr_t)addr & 4095; } static void semaphore_resolve(int i915) { const uint32_t SEMAPHORE_ADDR = 64 << 10; uint32_t semaphore, outer, inner, *sema; unsigned int engine; /* * Userspace may submit batches that wait upon unresolved * semaphores. Ideally, we want to put those blocking batches * to the back of the execution queue if we have something else * that is ready to run right away. This test exploits a failure * to reorder batches around a blocking semaphore by submitting * the release of that semaphore from a later context. */ igt_require(gem_scheduler_has_preemption(i915)); igt_require(intel_get_drm_devid(i915) >= 8); /* for MI_SEMAPHORE_WAIT */ outer = gem_context_create(i915); inner = gem_context_create(i915); semaphore = gem_create(i915, 4096); sema = gem_mmap__wc(i915, semaphore, 0, 4096, PROT_WRITE); for_each_physical_engine(i915, engine) { struct drm_i915_gem_exec_object2 obj[3]; struct drm_i915_gem_execbuffer2 eb; uint32_t handle, cancel; uint32_t *cs, *map; igt_spin_t *spin; int64_t poke = 1; if (!gem_can_store_dword(i915, engine)) continue; spin = __igt_spin_new(i915, .engine = engine); igt_spin_end(spin); /* we just want its address for later */ gem_sync(i915, spin->handle); igt_spin_reset(spin); handle = gem_create(i915, 4096); cs = map = gem_mmap__cpu(i915, handle, 0, 4096, PROT_WRITE); /* Set semaphore initially to 1 for polling and signaling */ *cs++ = MI_STORE_DWORD_IMM; *cs++ = SEMAPHORE_ADDR; *cs++ = 0; *cs++ = 1; /* Wait until another batch writes to our semaphore */ *cs++ = MI_SEMAPHORE_WAIT | MI_SEMAPHORE_POLL | MI_SEMAPHORE_SAD_EQ_SDD | (4 - 2); *cs++ = 0; *cs++ = SEMAPHORE_ADDR; *cs++ = 0; /* Then cancel the spinner */ *cs++ = MI_STORE_DWORD_IMM; *cs++ = spin->obj[IGT_SPIN_BATCH].offset + offset_in_page(spin->condition); *cs++ = 0; *cs++ = MI_BATCH_BUFFER_END; *cs++ = MI_BATCH_BUFFER_END; munmap(map, 4096); memset(&eb, 0, sizeof(eb)); /* First up is our spinning semaphore */ memset(obj, 0, sizeof(obj)); obj[0] = spin->obj[IGT_SPIN_BATCH]; obj[1].handle = semaphore; obj[1].offset = SEMAPHORE_ADDR; obj[1].flags = EXEC_OBJECT_PINNED; obj[2].handle = handle; eb.buffer_count = 3; eb.buffers_ptr = to_user_pointer(obj); eb.rsvd1 = outer; gem_execbuf(i915, &eb); /* Then add the GPU hang intermediatory */ memset(obj, 0, sizeof(obj)); obj[0].handle = handle; obj[0].flags = EXEC_OBJECT_WRITE; /* always after semaphore */ obj[1] = spin->obj[IGT_SPIN_BATCH]; eb.buffer_count = 2; eb.rsvd1 = 0; gem_execbuf(i915, &eb); while (READ_ONCE(*sema) == 0) ; /* Now the semaphore is spinning, cancel it */ cancel = gem_create(i915, 4096); cs = map = gem_mmap__cpu(i915, cancel, 0, 4096, PROT_WRITE); *cs++ = MI_STORE_DWORD_IMM; *cs++ = SEMAPHORE_ADDR; *cs++ = 0; *cs++ = 0; *cs++ = MI_BATCH_BUFFER_END; munmap(map, 4096); memset(obj, 0, sizeof(obj)); obj[0].handle = semaphore; obj[0].offset = SEMAPHORE_ADDR; obj[0].flags = EXEC_OBJECT_PINNED; obj[1].handle = cancel; eb.buffer_count = 2; eb.rsvd1 = inner; gem_execbuf(i915, &eb); gem_wait(i915, cancel, &poke); /* match sync's WAIT_PRIORITY */ gem_close(i915, cancel); gem_sync(i915, handle); /* To hang unless cancel runs! */ gem_close(i915, handle); igt_spin_free(i915, spin); igt_assert_eq(*sema, 0); } munmap(sema, 4096); gem_close(i915, semaphore); gem_context_destroy(i915, inner); gem_context_destroy(i915, outer); } static void semaphore_noskip(int i915) { const int gen = intel_gen(intel_get_drm_devid(i915)); unsigned int engine, other; uint32_t ctx; igt_require(gen >= 6); /* MI_STORE_DWORD_IMM convenience */ ctx = gem_context_create(i915); for_each_physical_engine(i915, engine) { for_each_physical_engine(i915, other) { struct drm_i915_gem_exec_object2 obj[3]; struct drm_i915_gem_execbuffer2 eb; uint32_t handle, *cs, *map; igt_spin_t *chain, *spin; if (other == engine || !gem_can_store_dword(i915, other)) continue; chain = __igt_spin_new(i915, .engine = engine); spin = __igt_spin_new(i915, .engine = other); igt_spin_end(spin); /* we just want its address for later */ gem_sync(i915, spin->handle); igt_spin_reset(spin); handle = gem_create(i915, 4096); cs = map = gem_mmap__cpu(i915, handle, 0, 4096, PROT_WRITE); /* Cancel the following spinner */ *cs++ = MI_STORE_DWORD_IMM; if (gen >= 8) { *cs++ = spin->obj[IGT_SPIN_BATCH].offset + offset_in_page(spin->condition); *cs++ = 0; } else { *cs++ = 0; *cs++ = spin->obj[IGT_SPIN_BATCH].offset + offset_in_page(spin->condition); } *cs++ = MI_BATCH_BUFFER_END; *cs++ = MI_BATCH_BUFFER_END; munmap(map, 4096); /* port0: implicit semaphore from engine */ memset(obj, 0, sizeof(obj)); obj[0] = chain->obj[IGT_SPIN_BATCH]; obj[0].flags |= EXEC_OBJECT_WRITE; obj[1] = spin->obj[IGT_SPIN_BATCH]; obj[2].handle = handle; memset(&eb, 0, sizeof(eb)); eb.buffer_count = 3; eb.buffers_ptr = to_user_pointer(obj); eb.rsvd1 = ctx; eb.flags = other; gem_execbuf(i915, &eb); /* port1: dependency chain from port0 */ memset(obj, 0, sizeof(obj)); obj[0].handle = handle; obj[0].flags = EXEC_OBJECT_WRITE; obj[1] = spin->obj[IGT_SPIN_BATCH]; memset(&eb, 0, sizeof(eb)); eb.buffer_count = 2; eb.buffers_ptr = to_user_pointer(obj); eb.flags = other; gem_execbuf(i915, &eb); igt_spin_set_timeout(chain, NSEC_PER_SEC / 100); gem_sync(i915, spin->handle); /* To hang unless cancel runs! */ gem_close(i915, handle); igt_spin_free(i915, spin); igt_spin_free(i915, chain); } } gem_context_destroy(i915, ctx); } static void reorder(int fd, unsigned ring, unsigned flags) #define EQUAL 1 { IGT_CORK_HANDLE(cork); uint32_t scratch, plug; uint32_t result; uint32_t ctx[2]; ctx[LO] = gem_context_create(fd); gem_context_set_priority(fd, ctx[LO], MIN_PRIO); ctx[HI] = gem_context_create(fd); gem_context_set_priority(fd, ctx[HI], flags & EQUAL ? MIN_PRIO : 0); scratch = gem_create(fd, 4096); plug = igt_cork_plug(&cork, fd); /* We expect the high priority context to be executed first, and * so the final result will be value from the low priority context. */ store_dword(fd, ctx[LO], ring, scratch, 0, ctx[LO], plug, 0); store_dword(fd, ctx[HI], ring, scratch, 0, ctx[HI], plug, 0); unplug_show_queue(fd, &cork, ring); gem_close(fd, plug); gem_context_destroy(fd, ctx[LO]); gem_context_destroy(fd, ctx[HI]); result = __sync_read_u32(fd, scratch, 0); gem_close(fd, scratch); if (flags & EQUAL) /* equal priority, result will be fifo */ igt_assert_eq_u32(result, ctx[HI]); else igt_assert_eq_u32(result, ctx[LO]); } static void promotion(int fd, unsigned ring) { IGT_CORK_HANDLE(cork); uint32_t result, dep; uint32_t result_read, dep_read; uint32_t ctx[3]; uint32_t plug; ctx[LO] = gem_context_create(fd); gem_context_set_priority(fd, ctx[LO], MIN_PRIO); ctx[HI] = gem_context_create(fd); gem_context_set_priority(fd, ctx[HI], 0); ctx[NOISE] = gem_context_create(fd); gem_context_set_priority(fd, ctx[NOISE], MIN_PRIO/2); result = gem_create(fd, 4096); dep = gem_create(fd, 4096); plug = igt_cork_plug(&cork, fd); /* Expect that HI promotes LO, so the order will be LO, HI, NOISE. * * fifo would be NOISE, LO, HI. * strict priority would be HI, NOISE, LO */ store_dword(fd, ctx[NOISE], ring, result, 0, ctx[NOISE], plug, 0); store_dword(fd, ctx[LO], ring, result, 0, ctx[LO], plug, 0); /* link LO <-> HI via a dependency on another buffer */ store_dword(fd, ctx[LO], ring, dep, 0, ctx[LO], 0, I915_GEM_DOMAIN_INSTRUCTION); store_dword(fd, ctx[HI], ring, dep, 0, ctx[HI], 0, 0); store_dword(fd, ctx[HI], ring, result, 0, ctx[HI], 0, 0); unplug_show_queue(fd, &cork, ring); gem_close(fd, plug); gem_context_destroy(fd, ctx[NOISE]); gem_context_destroy(fd, ctx[LO]); gem_context_destroy(fd, ctx[HI]); dep_read = __sync_read_u32(fd, dep, 0); gem_close(fd, dep); result_read = __sync_read_u32(fd, result, 0); gem_close(fd, result); igt_assert_eq_u32(dep_read, ctx[HI]); igt_assert_eq_u32(result_read, ctx[NOISE]); } #define NEW_CTX (0x1 << 0) #define HANG_LP (0x1 << 1) static void preempt(int fd, unsigned ring, unsigned flags) { uint32_t result = gem_create(fd, 4096); uint32_t result_read; igt_spin_t *spin[MAX_ELSP_QLEN]; uint32_t ctx[2]; igt_hang_t hang; ctx[LO] = gem_context_create(fd); gem_context_set_priority(fd, ctx[LO], MIN_PRIO); ctx[HI] = gem_context_create(fd); gem_context_set_priority(fd, ctx[HI], MAX_PRIO); if (flags & HANG_LP) hang = igt_hang_ctx(fd, ctx[LO], ring, 0); for (int n = 0; n < ARRAY_SIZE(spin); n++) { if (flags & NEW_CTX) { gem_context_destroy(fd, ctx[LO]); ctx[LO] = gem_context_create(fd); gem_context_set_priority(fd, ctx[LO], MIN_PRIO); } spin[n] = __igt_spin_new(fd, .ctx = ctx[LO], .engine = ring); igt_debug("spin[%d].handle=%d\n", n, spin[n]->handle); store_dword(fd, ctx[HI], ring, result, 0, n + 1, 0, I915_GEM_DOMAIN_RENDER); result_read = __sync_read_u32(fd, result, 0); igt_assert_eq_u32(result_read, n + 1); igt_assert(gem_bo_busy(fd, spin[0]->handle)); } for (int n = 0; n < ARRAY_SIZE(spin); n++) igt_spin_free(fd, spin[n]); if (flags & HANG_LP) igt_post_hang_ring(fd, hang); gem_context_destroy(fd, ctx[LO]); gem_context_destroy(fd, ctx[HI]); gem_close(fd, result); } #define CHAIN 0x1 #define CONTEXTS 0x2 static igt_spin_t *__noise(int fd, uint32_t ctx, int prio, igt_spin_t *spin) { unsigned other; gem_context_set_priority(fd, ctx, prio); for_each_physical_engine(fd, other) { if (spin == NULL) { spin = __igt_spin_new(fd, .ctx = ctx, .engine = other); } else { struct drm_i915_gem_execbuffer2 eb = { .buffer_count = 1, .buffers_ptr = to_user_pointer(&spin->obj[IGT_SPIN_BATCH]), .rsvd1 = ctx, .flags = other, }; gem_execbuf(fd, &eb); } } return spin; } static void __preempt_other(int fd, uint32_t *ctx, unsigned int target, unsigned int primary, unsigned flags) { uint32_t result = gem_create(fd, 4096); uint32_t result_read[4096 / sizeof(uint32_t)]; unsigned int n, i, other; n = 0; store_dword(fd, ctx[LO], primary, result, (n + 1)*sizeof(uint32_t), n + 1, 0, I915_GEM_DOMAIN_RENDER); n++; if (flags & CHAIN) { for_each_physical_engine(fd, other) { store_dword(fd, ctx[LO], other, result, (n + 1)*sizeof(uint32_t), n + 1, 0, I915_GEM_DOMAIN_RENDER); n++; } } store_dword(fd, ctx[HI], target, result, (n + 1)*sizeof(uint32_t), n + 1, 0, I915_GEM_DOMAIN_RENDER); igt_debugfs_dump(fd, "i915_engine_info"); gem_set_domain(fd, result, I915_GEM_DOMAIN_GTT, 0); n++; __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); for (i = 0; i <= n; i++) igt_assert_eq_u32(result_read[i], i); gem_close(fd, result); } static void preempt_other(int fd, unsigned ring, unsigned int flags) { unsigned int primary; igt_spin_t *spin = NULL; uint32_t ctx[3]; /* On each engine, insert * [NOISE] spinner, * [LOW] write * * Then on our target engine do a [HIGH] write which should then * prompt its dependent LOW writes in front of the spinner on * each engine. The purpose of this test is to check that preemption * can cross engines. */ ctx[LO] = gem_context_create(fd); gem_context_set_priority(fd, ctx[LO], MIN_PRIO); ctx[NOISE] = gem_context_create(fd); spin = __noise(fd, ctx[NOISE], 0, NULL); ctx[HI] = gem_context_create(fd); gem_context_set_priority(fd, ctx[HI], MAX_PRIO); for_each_physical_engine(fd, primary) { igt_debug("Primary engine: %s\n", e__->name); __preempt_other(fd, ctx, ring, primary, flags); } igt_assert(gem_bo_busy(fd, spin->handle)); igt_spin_free(fd, spin); gem_context_destroy(fd, ctx[LO]); gem_context_destroy(fd, ctx[NOISE]); gem_context_destroy(fd, ctx[HI]); } static void __preempt_queue(int fd, unsigned target, unsigned primary, unsigned depth, unsigned flags) { uint32_t result = gem_create(fd, 4096); uint32_t result_read[4096 / sizeof(uint32_t)]; igt_spin_t *above = NULL, *below = NULL; unsigned int other, n, i; int prio = MAX_PRIO; uint32_t ctx[3] = { gem_context_create(fd), gem_context_create(fd), gem_context_create(fd), }; for (n = 0; n < depth; n++) { if (flags & CONTEXTS) { gem_context_destroy(fd, ctx[NOISE]); ctx[NOISE] = gem_context_create(fd); } above = __noise(fd, ctx[NOISE], prio--, above); } gem_context_set_priority(fd, ctx[HI], prio--); for (; n < MAX_ELSP_QLEN; n++) { if (flags & CONTEXTS) { gem_context_destroy(fd, ctx[NOISE]); ctx[NOISE] = gem_context_create(fd); } below = __noise(fd, ctx[NOISE], prio--, below); } gem_context_set_priority(fd, ctx[LO], prio--); n = 0; store_dword(fd, ctx[LO], primary, result, (n + 1)*sizeof(uint32_t), n + 1, 0, I915_GEM_DOMAIN_RENDER); n++; if (flags & CHAIN) { for_each_physical_engine(fd, other) { store_dword(fd, ctx[LO], other, result, (n + 1)*sizeof(uint32_t), n + 1, 0, I915_GEM_DOMAIN_RENDER); n++; } } store_dword(fd, ctx[HI], target, result, (n + 1)*sizeof(uint32_t), n + 1, 0, I915_GEM_DOMAIN_RENDER); igt_debugfs_dump(fd, "i915_engine_info"); if (above) { igt_assert(gem_bo_busy(fd, above->handle)); igt_spin_free(fd, above); } gem_set_domain(fd, result, I915_GEM_DOMAIN_GTT, 0); __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); n++; for (i = 0; i <= n; i++) igt_assert_eq_u32(result_read[i], i); if (below) { igt_assert(gem_bo_busy(fd, below->handle)); igt_spin_free(fd, below); } gem_context_destroy(fd, ctx[LO]); gem_context_destroy(fd, ctx[NOISE]); gem_context_destroy(fd, ctx[HI]); gem_close(fd, result); } static void preempt_queue(int fd, unsigned ring, unsigned int flags) { unsigned other; for_each_physical_engine(fd, other) { for (unsigned depth = 0; depth <= MAX_ELSP_QLEN; depth++) __preempt_queue(fd, ring, other, depth, flags); } } static void preempt_self(int fd, unsigned ring) { uint32_t result = gem_create(fd, 4096); uint32_t result_read[4096 / sizeof(uint32_t)]; igt_spin_t *spin[MAX_ELSP_QLEN]; unsigned int other; unsigned int n, i; uint32_t ctx[3]; /* On each engine, insert * [NOISE] spinner, * [self/LOW] write * * Then on our target engine do a [self/HIGH] write which should then * preempt its own lower priority task on any engine. */ ctx[NOISE] = gem_context_create(fd); ctx[HI] = gem_context_create(fd); n = 0; gem_context_set_priority(fd, ctx[HI], MIN_PRIO); for_each_physical_engine(fd, other) { spin[n] = __igt_spin_new(fd, .ctx = ctx[NOISE], .engine = other); store_dword(fd, ctx[HI], other, result, (n + 1)*sizeof(uint32_t), n + 1, 0, I915_GEM_DOMAIN_RENDER); n++; } gem_context_set_priority(fd, ctx[HI], MAX_PRIO); store_dword(fd, ctx[HI], ring, result, (n + 1)*sizeof(uint32_t), n + 1, 0, I915_GEM_DOMAIN_RENDER); gem_set_domain(fd, result, I915_GEM_DOMAIN_GTT, 0); for (i = 0; i < n; i++) { igt_assert(gem_bo_busy(fd, spin[i]->handle)); igt_spin_free(fd, spin[i]); } __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); n++; for (i = 0; i <= n; i++) igt_assert_eq_u32(result_read[i], i); gem_context_destroy(fd, ctx[NOISE]); gem_context_destroy(fd, ctx[HI]); gem_close(fd, result); } static void preemptive_hang(int fd, unsigned ring) { igt_spin_t *spin[MAX_ELSP_QLEN]; igt_hang_t hang; uint32_t ctx[2]; ctx[HI] = gem_context_create(fd); gem_context_set_priority(fd, ctx[HI], MAX_PRIO); for (int n = 0; n < ARRAY_SIZE(spin); n++) { ctx[LO] = gem_context_create(fd); gem_context_set_priority(fd, ctx[LO], MIN_PRIO); spin[n] = __igt_spin_new(fd, .ctx = ctx[LO], .engine = ring); gem_context_destroy(fd, ctx[LO]); } hang = igt_hang_ctx(fd, ctx[HI], ring, 0); igt_post_hang_ring(fd, hang); for (int n = 0; n < ARRAY_SIZE(spin); n++) { /* Current behavior is to execute requests in order of submission. * This is subject to change as the scheduler evolve. The test should * be updated to reflect such changes. */ igt_assert(gem_bo_busy(fd, spin[n]->handle)); igt_spin_free(fd, spin[n]); } gem_context_destroy(fd, ctx[HI]); } static void deep(int fd, unsigned ring) { #define XS 8 const unsigned int max_req = MAX_PRIO - MIN_PRIO; const unsigned size = ALIGN(4*max_req, 4096); struct timespec tv = {}; IGT_CORK_HANDLE(cork); unsigned int nreq; uint32_t plug; uint32_t result, dep[XS]; uint32_t read_buf[size / sizeof(uint32_t)]; uint32_t expected = 0; uint32_t *ctx; int dep_nreq; int n; ctx = malloc(sizeof(*ctx) * MAX_CONTEXTS); for (n = 0; n < MAX_CONTEXTS; n++) { ctx[n] = gem_context_create(fd); } nreq = gem_measure_ring_inflight(fd, ring, 0) / (4 * XS) * MAX_CONTEXTS; if (nreq > max_req) nreq = max_req; igt_info("Using %d requests (prio range %d)\n", nreq, max_req); result = gem_create(fd, size); for (int m = 0; m < XS; m ++) dep[m] = gem_create(fd, size); /* Bind all surfaces and contexts before starting the timeout. */ { struct drm_i915_gem_exec_object2 obj[XS + 2]; struct drm_i915_gem_execbuffer2 execbuf; const uint32_t bbe = MI_BATCH_BUFFER_END; memset(obj, 0, sizeof(obj)); for (n = 0; n < XS; n++) obj[n].handle = dep[n]; obj[XS].handle = result; obj[XS+1].handle = gem_create(fd, 4096); gem_write(fd, obj[XS+1].handle, 0, &bbe, sizeof(bbe)); memset(&execbuf, 0, sizeof(execbuf)); execbuf.buffers_ptr = to_user_pointer(obj); execbuf.buffer_count = XS + 2; execbuf.flags = ring; for (n = 0; n < MAX_CONTEXTS; n++) { execbuf.rsvd1 = ctx[n]; gem_execbuf(fd, &execbuf); } gem_close(fd, obj[XS+1].handle); gem_sync(fd, result); } plug = igt_cork_plug(&cork, fd); /* Create a deep dependency chain, with a few branches */ for (n = 0; n < nreq && igt_seconds_elapsed(&tv) < 2; n++) { const int gen = intel_gen(intel_get_drm_devid(fd)); struct drm_i915_gem_exec_object2 obj[3]; struct drm_i915_gem_relocation_entry reloc; struct drm_i915_gem_execbuffer2 eb = { .buffers_ptr = to_user_pointer(obj), .buffer_count = 3, .flags = ring | (gen < 6 ? I915_EXEC_SECURE : 0), .rsvd1 = ctx[n % MAX_CONTEXTS], }; uint32_t batch[16]; int i; memset(obj, 0, sizeof(obj)); obj[0].handle = plug; memset(&reloc, 0, sizeof(reloc)); reloc.presumed_offset = 0; reloc.offset = sizeof(uint32_t); reloc.delta = sizeof(uint32_t) * n; reloc.read_domains = I915_GEM_DOMAIN_RENDER; reloc.write_domain = I915_GEM_DOMAIN_RENDER; obj[2].handle = gem_create(fd, 4096); obj[2].relocs_ptr = to_user_pointer(&reloc); obj[2].relocation_count = 1; i = 0; batch[i] = MI_STORE_DWORD_IMM | (gen < 6 ? 1 << 22 : 0); if (gen >= 8) { batch[++i] = reloc.delta; batch[++i] = 0; } else if (gen >= 4) { batch[++i] = 0; batch[++i] = reloc.delta; reloc.offset += sizeof(uint32_t); } else { batch[i]--; batch[++i] = reloc.delta; } batch[++i] = eb.rsvd1; batch[++i] = MI_BATCH_BUFFER_END; gem_write(fd, obj[2].handle, 0, batch, sizeof(batch)); gem_context_set_priority(fd, eb.rsvd1, MAX_PRIO - nreq + n); for (int m = 0; m < XS; m++) { obj[1].handle = dep[m]; reloc.target_handle = obj[1].handle; gem_execbuf(fd, &eb); } gem_close(fd, obj[2].handle); } igt_info("First deptree: %d requests [%.3fs]\n", n * XS, 1e-9*igt_nsec_elapsed(&tv)); dep_nreq = n; for (n = 0; n < nreq && igt_seconds_elapsed(&tv) < 4; n++) { uint32_t context = ctx[n % MAX_CONTEXTS]; gem_context_set_priority(fd, context, MAX_PRIO - nreq + n); for (int m = 0; m < XS; m++) { store_dword(fd, context, ring, result, 4*n, context, dep[m], 0); store_dword(fd, context, ring, result, 4*m, context, 0, I915_GEM_DOMAIN_INSTRUCTION); } expected = context; } igt_info("Second deptree: %d requests [%.3fs]\n", n * XS, 1e-9*igt_nsec_elapsed(&tv)); unplug_show_queue(fd, &cork, ring); gem_close(fd, plug); igt_require(expected); /* too slow */ for (n = 0; n < MAX_CONTEXTS; n++) gem_context_destroy(fd, ctx[n]); for (int m = 0; m < XS; m++) { __sync_read_u32_count(fd, dep[m], read_buf, sizeof(read_buf)); gem_close(fd, dep[m]); for (n = 0; n < dep_nreq; n++) igt_assert_eq_u32(read_buf[n], ctx[n % MAX_CONTEXTS]); } __sync_read_u32_count(fd, result, read_buf, sizeof(read_buf)); gem_close(fd, result); /* No reordering due to PI on all contexts because of the common dep */ for (int m = 0; m < XS; m++) igt_assert_eq_u32(read_buf[m], expected); free(ctx); #undef XS } static void alarm_handler(int sig) { } static int __execbuf(int fd, struct drm_i915_gem_execbuffer2 *execbuf) { int err = 0; if (ioctl(fd, DRM_IOCTL_I915_GEM_EXECBUFFER2, execbuf)) err = -errno; return err; } static void wide(int fd, unsigned ring) { struct timespec tv = {}; unsigned int ring_size = gem_measure_ring_inflight(fd, ring, MEASURE_RING_NEW_CTX); IGT_CORK_HANDLE(cork); uint32_t plug; uint32_t result; uint32_t result_read[MAX_CONTEXTS]; uint32_t *ctx; unsigned int count; ctx = malloc(sizeof(*ctx)*MAX_CONTEXTS); for (int n = 0; n < MAX_CONTEXTS; n++) ctx[n] = gem_context_create(fd); result = gem_create(fd, 4*MAX_CONTEXTS); plug = igt_cork_plug(&cork, fd); /* Lots of in-order requests, plugged and submitted simultaneously */ for (count = 0; igt_seconds_elapsed(&tv) < 5 && count < ring_size; count++) { for (int n = 0; n < MAX_CONTEXTS; n++) { store_dword(fd, ctx[n], ring, result, 4*n, ctx[n], plug, I915_GEM_DOMAIN_INSTRUCTION); } } igt_info("Submitted %d requests over %d contexts in %.1fms\n", count, MAX_CONTEXTS, igt_nsec_elapsed(&tv) * 1e-6); unplug_show_queue(fd, &cork, ring); gem_close(fd, plug); for (int n = 0; n < MAX_CONTEXTS; n++) gem_context_destroy(fd, ctx[n]); __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); for (int n = 0; n < MAX_CONTEXTS; n++) igt_assert_eq_u32(result_read[n], ctx[n]); gem_close(fd, result); free(ctx); } static void reorder_wide(int fd, unsigned ring) { const int gen = intel_gen(intel_get_drm_devid(fd)); struct drm_i915_gem_relocation_entry reloc; struct drm_i915_gem_exec_object2 obj[3]; struct drm_i915_gem_execbuffer2 execbuf; struct timespec tv = {}; unsigned int ring_size = gem_measure_ring_inflight(fd, ring, MEASURE_RING_NEW_CTX); IGT_CORK_HANDLE(cork); uint32_t result, target, plug; uint32_t result_read[1024]; uint32_t *expected; result = gem_create(fd, 4096); target = gem_create(fd, 4096); plug = igt_cork_plug(&cork, fd); expected = gem_mmap__cpu(fd, target, 0, 4096, PROT_WRITE); gem_set_domain(fd, target, I915_GEM_DOMAIN_CPU, I915_GEM_DOMAIN_CPU); memset(obj, 0, sizeof(obj)); obj[0].handle = plug; obj[1].handle = result; obj[2].relocs_ptr = to_user_pointer(&reloc); obj[2].relocation_count = 1; memset(&reloc, 0, sizeof(reloc)); reloc.target_handle = result; reloc.read_domains = I915_GEM_DOMAIN_INSTRUCTION; reloc.write_domain = 0; /* lies */ memset(&execbuf, 0, sizeof(execbuf)); execbuf.buffers_ptr = to_user_pointer(obj); execbuf.buffer_count = 3; execbuf.flags = ring; if (gen < 6) execbuf.flags |= I915_EXEC_SECURE; for (int n = MIN_PRIO, x = 1; igt_seconds_elapsed(&tv) < 5 && n <= MAX_PRIO; n++, x++) { unsigned int sz = ALIGN(ring_size * 64, 4096); uint32_t *batch; execbuf.rsvd1 = gem_context_create(fd); gem_context_set_priority(fd, execbuf.rsvd1, n); obj[2].handle = gem_create(fd, sz); batch = gem_mmap__gtt(fd, obj[2].handle, sz, PROT_WRITE); gem_set_domain(fd, obj[2].handle, I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT); for (int m = 0; m < ring_size; m++) { uint64_t addr; int idx = hars_petruska_f54_1_random_unsafe_max(1024); int i; execbuf.batch_start_offset = m * 64; reloc.offset = execbuf.batch_start_offset + sizeof(uint32_t); reloc.delta = idx * sizeof(uint32_t); addr = reloc.presumed_offset + reloc.delta; i = execbuf.batch_start_offset / sizeof(uint32_t); batch[i] = MI_STORE_DWORD_IMM | (gen < 6 ? 1 << 22 : 0); if (gen >= 8) { batch[++i] = addr; batch[++i] = addr >> 32; } else if (gen >= 4) { batch[++i] = 0; batch[++i] = addr; reloc.offset += sizeof(uint32_t); } else { batch[i]--; batch[++i] = addr; } batch[++i] = x; batch[++i] = MI_BATCH_BUFFER_END; if (!expected[idx]) expected[idx] = x; gem_execbuf(fd, &execbuf); } munmap(batch, sz); gem_close(fd, obj[2].handle); gem_context_destroy(fd, execbuf.rsvd1); } unplug_show_queue(fd, &cork, ring); gem_close(fd, plug); __sync_read_u32_count(fd, result, result_read, sizeof(result_read)); for (int n = 0; n < 1024; n++) igt_assert_eq_u32(result_read[n], expected[n]); munmap(expected, 4096); gem_close(fd, result); gem_close(fd, target); } static void bind_to_cpu(int cpu) { const int ncpus = sysconf(_SC_NPROCESSORS_ONLN); struct sched_param rt = {.sched_priority = 99 }; cpu_set_t allowed; igt_assert(sched_setscheduler(getpid(), SCHED_RR | SCHED_RESET_ON_FORK, &rt) == 0); CPU_ZERO(&allowed); CPU_SET(cpu % ncpus, &allowed); igt_assert(sched_setaffinity(getpid(), sizeof(cpu_set_t), &allowed) == 0); } static void test_pi_ringfull(int fd, unsigned int engine) { const uint32_t bbe = MI_BATCH_BUFFER_END; struct sigaction sa = { .sa_handler = alarm_handler }; struct drm_i915_gem_execbuffer2 execbuf; struct drm_i915_gem_exec_object2 obj[2]; unsigned int last, count; struct itimerval itv; IGT_CORK_HANDLE(c); uint32_t vip; bool *result; result = mmap(NULL, 4096, PROT_WRITE, MAP_SHARED | MAP_ANON, -1, 0); igt_assert(result != MAP_FAILED); memset(&execbuf, 0, sizeof(execbuf)); memset(&obj, 0, sizeof(obj)); obj[1].handle = gem_create(fd, 4096); gem_write(fd, obj[1].handle, 0, &bbe, sizeof(bbe)); execbuf.buffers_ptr = to_user_pointer(&obj[1]); execbuf.buffer_count = 1; execbuf.flags = engine; /* Warm up both (hi/lo) contexts */ execbuf.rsvd1 = gem_context_create(fd); gem_context_set_priority(fd, execbuf.rsvd1, MAX_PRIO); gem_execbuf(fd, &execbuf); gem_sync(fd, obj[1].handle); vip = execbuf.rsvd1; execbuf.rsvd1 = gem_context_create(fd); gem_context_set_priority(fd, execbuf.rsvd1, MIN_PRIO); gem_execbuf(fd, &execbuf); gem_sync(fd, obj[1].handle); /* Fill the low-priority ring */ obj[0].handle = igt_cork_plug(&c, fd); execbuf.buffers_ptr = to_user_pointer(obj); execbuf.buffer_count = 2; sigaction(SIGALRM, &sa, NULL); itv.it_interval.tv_sec = 0; itv.it_interval.tv_usec = 1000; itv.it_value.tv_sec = 0; itv.it_value.tv_usec = 10000; setitimer(ITIMER_REAL, &itv, NULL); last = -1; count = 0; do { if (__execbuf(fd, &execbuf) == 0) { count++; continue; } if (last == count) break; last = count; } while (1); igt_debug("Filled low-priority ring with %d batches\n", count); memset(&itv, 0, sizeof(itv)); setitimer(ITIMER_REAL, &itv, NULL); execbuf.buffers_ptr = to_user_pointer(&obj[1]); execbuf.buffer_count = 1; /* both parent + child on the same cpu, only parent is RT */ bind_to_cpu(0); igt_fork(child, 1) { int err; result[0] = vip != execbuf.rsvd1; igt_debug("Waking parent\n"); kill(getppid(), SIGALRM); sched_yield(); result[1] = true; sigaction(SIGALRM, &sa, NULL); itv.it_value.tv_sec = 0; itv.it_value.tv_usec = 10000; setitimer(ITIMER_REAL, &itv, NULL); /* Since we are the high priority task, we expect to be * able to add ourselves to *our* ring without interruption. */ igt_debug("HP child executing\n"); execbuf.rsvd1 = vip; err = __execbuf(fd, &execbuf); igt_debug("HP execbuf returned %d\n", err); memset(&itv, 0, sizeof(itv)); setitimer(ITIMER_REAL, &itv, NULL); result[2] = err == 0; } /* Relinquish CPU just to allow child to create a context */ sleep(1); igt_assert_f(result[0], "HP context (child) not created\n"); igt_assert_f(!result[1], "Child released too early!\n"); /* Parent sleeps waiting for ringspace, releasing child */ itv.it_value.tv_sec = 0; itv.it_value.tv_usec = 50000; setitimer(ITIMER_REAL, &itv, NULL); igt_debug("LP parent executing\n"); igt_assert_eq(__execbuf(fd, &execbuf), -EINTR); igt_assert_f(result[1], "Child was not released!\n"); igt_assert_f(result[2], "High priority child unable to submit within 10ms\n"); igt_cork_unplug(&c); igt_waitchildren(); gem_context_destroy(fd, execbuf.rsvd1); gem_context_destroy(fd, vip); gem_close(fd, obj[1].handle); gem_close(fd, obj[0].handle); munmap(result, 4096); } static void measure_semaphore_power(int i915) { struct gpu_power power; unsigned int engine, signaler; igt_require(gpu_power_open(&power) == 0); for_each_physical_engine(i915, signaler) { struct gpu_power_sample s_spin[2]; struct gpu_power_sample s_sema[2]; double baseline, total; int64_t jiffie = 1; igt_spin_t *spin; spin = __igt_spin_new(i915, .engine = signaler, .flags = IGT_SPIN_POLL_RUN); gem_wait(i915, spin->handle, &jiffie); /* waitboost */ igt_spin_busywait_until_started(spin); gpu_power_read(&power, &s_spin[0]); usleep(100*1000); gpu_power_read(&power, &s_spin[1]); /* Add a waiter to each engine */ for_each_physical_engine(i915, engine) { igt_spin_t *sema; if (engine == signaler) continue; sema = __igt_spin_new(i915, .engine = engine, .dependency = spin->handle); igt_spin_free(i915, sema); } usleep(10); /* just give the tasklets a chance to run */ gpu_power_read(&power, &s_sema[0]); usleep(100*1000); gpu_power_read(&power, &s_sema[1]); igt_spin_free(i915, spin); baseline = gpu_power_W(&power, &s_spin[0], &s_spin[1]); total = gpu_power_W(&power, &s_sema[0], &s_sema[1]); igt_info("%s: %.1fmW + %.1fmW (total %1.fmW)\n", e__->name, 1e3 * baseline, 1e3 * (total - baseline), 1e3 * total); } gpu_power_close(&power); } igt_main { const struct intel_execution_engine *e; int fd = -1; igt_skip_on_simulation(); igt_fixture { fd = drm_open_driver_master(DRIVER_INTEL); gem_submission_print_method(fd); gem_scheduler_print_capability(fd); igt_require_gem(fd); gem_require_mmap_wc(fd); gem_require_contexts(fd); igt_fork_hang_detector(fd); } igt_subtest_group { for (e = intel_execution_engines; e->name; e++) { /* default exec-id is purely symbolic */ if (e->exec_id == 0) continue; igt_subtest_f("fifo-%s", e->name) { igt_require(gem_ring_has_physical_engine(fd, e->exec_id | e->flags)); igt_require(gem_can_store_dword(fd, e->exec_id | e->flags)); fifo(fd, e->exec_id | e->flags); } igt_subtest_f("independent-%s", e->name) { igt_require(gem_ring_has_physical_engine(fd, e->exec_id | e->flags)); igt_require(gem_can_store_dword(fd, e->exec_id | e->flags)); independent(fd, e->exec_id | e->flags); } } } igt_subtest_group { igt_fixture { igt_require(gem_scheduler_enabled(fd)); igt_require(gem_scheduler_has_ctx_priority(fd)); } igt_subtest("semaphore-user") semaphore_userlock(fd); igt_subtest("semaphore-codependency") semaphore_codependency(fd); igt_subtest("semaphore-resolve") semaphore_resolve(fd); igt_subtest("semaphore-noskip") semaphore_noskip(fd); igt_subtest("smoketest-all") smoketest(fd, ALL_ENGINES, 30); for (e = intel_execution_engines; e->name; e++) { if (e->exec_id == 0) continue; igt_subtest_group { igt_fixture { igt_require(gem_ring_has_physical_engine(fd, e->exec_id | e->flags)); igt_require(gem_can_store_dword(fd, e->exec_id | e->flags)); } igt_subtest_f("in-order-%s", e->name) reorder(fd, e->exec_id | e->flags, EQUAL); igt_subtest_f("out-order-%s", e->name) reorder(fd, e->exec_id | e->flags, 0); igt_subtest_f("promotion-%s", e->name) promotion(fd, e->exec_id | e->flags); igt_subtest_group { igt_fixture { igt_require(gem_scheduler_has_preemption(fd)); } igt_subtest_f("preempt-%s", e->name) preempt(fd, e->exec_id | e->flags, 0); igt_subtest_f("preempt-contexts-%s", e->name) preempt(fd, e->exec_id | e->flags, NEW_CTX); igt_subtest_f("preempt-self-%s", e->name) preempt_self(fd, e->exec_id | e->flags); igt_subtest_f("preempt-other-%s", e->name) preempt_other(fd, e->exec_id | e->flags, 0); igt_subtest_f("preempt-other-chain-%s", e->name) preempt_other(fd, e->exec_id | e->flags, CHAIN); igt_subtest_f("preempt-queue-%s", e->name) preempt_queue(fd, e->exec_id | e->flags, 0); igt_subtest_f("preempt-queue-chain-%s", e->name) preempt_queue(fd, e->exec_id | e->flags, CHAIN); igt_subtest_f("preempt-queue-contexts-%s", e->name) preempt_queue(fd, e->exec_id | e->flags, CONTEXTS); igt_subtest_f("preempt-queue-contexts-chain-%s", e->name) preempt_queue(fd, e->exec_id | e->flags, CONTEXTS | CHAIN); igt_subtest_group { igt_hang_t hang; igt_fixture { igt_stop_hang_detector(); hang = igt_allow_hang(fd, 0, 0); } igt_subtest_f("preempt-hang-%s", e->name) { preempt(fd, e->exec_id | e->flags, NEW_CTX | HANG_LP); } igt_subtest_f("preemptive-hang-%s", e->name) preemptive_hang(fd, e->exec_id | e->flags); igt_fixture { igt_disallow_hang(fd, hang); igt_fork_hang_detector(fd); } } } igt_subtest_f("deep-%s", e->name) deep(fd, e->exec_id | e->flags); igt_subtest_f("wide-%s", e->name) wide(fd, e->exec_id | e->flags); igt_subtest_f("reorder-wide-%s", e->name) reorder_wide(fd, e->exec_id | e->flags); igt_subtest_f("smoketest-%s", e->name) smoketest(fd, e->exec_id | e->flags, 5); } } } igt_subtest_group { igt_fixture { igt_require(gem_scheduler_enabled(fd)); igt_require(gem_scheduler_has_ctx_priority(fd)); } for (e = intel_execution_engines; e->name; e++) { if (e->exec_id == 0) continue; igt_subtest_group { igt_fixture { igt_require(gem_ring_has_physical_engine(fd, e->exec_id | e->flags)); igt_require(gem_scheduler_has_preemption(fd)); } igt_subtest_f("pi-ringfull-%s", e->name) test_pi_ringfull(fd, e->exec_id | e->flags); } } } igt_subtest_group { igt_fixture { igt_require(gem_scheduler_enabled(fd)); igt_require(gem_scheduler_has_semaphores(fd)); } igt_subtest("semaphore-power") measure_semaphore_power(fd); } igt_fixture { igt_stop_hang_detector(); close(fd); } }