/* * Copyright © 2011 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. * * Authors: * Chris Wilson * */ #include "igt.h" #include #include #include #include #include #include #include #include #include #include #include #include #include #include "drm.h" #define LOCAL_I915_EXEC_NO_RELOC (1<<11) #define LOCAL_I915_EXEC_HANDLE_LUT (1<<12) #define INTERRUPTIBLE 0x1 #define QUEUE 0x2 static double elapsed(const struct timespec *start, const struct timespec *end) { return ((end->tv_sec - start->tv_sec) + (end->tv_nsec - start->tv_nsec)*1e-9); } static int measure_qlen(int fd, struct drm_i915_gem_execbuffer2 *execbuf, const struct intel_engine_data *engines, int timeout) { const struct drm_i915_gem_exec_object2 * const obj = (struct drm_i915_gem_exec_object2 *)(uintptr_t)execbuf->buffers_ptr; uint32_t ctx[64]; int min = INT_MAX, max = 0; for (int i = 0; i < ARRAY_SIZE(ctx); i++) { ctx[i] = gem_context_create(fd); gem_context_set_all_engines(fd, ctx[i]); } for (unsigned int n = 0; n < engines->nengines; n++) { uint64_t saved = execbuf->flags; struct timespec tv = {}; int q; execbuf->flags |= engines->engines[n].flags; for (int i = 0; i < ARRAY_SIZE(ctx); i++) { execbuf->rsvd1 = ctx[i]; gem_execbuf(fd, execbuf); } gem_sync(fd, obj->handle); igt_nsec_elapsed(&tv); for (int i = 0; i < ARRAY_SIZE(ctx); i++) { execbuf->rsvd1 = ctx[i]; gem_execbuf(fd, execbuf); } gem_sync(fd, obj->handle); /* * Be conservative and aim not to overshoot timeout, so scale * down by 8 for hopefully a max of 12.5% error. */ q = ARRAY_SIZE(ctx) * timeout * 1e9 / igt_nsec_elapsed(&tv) / 8 + 1; if (q < min) min = q; if (q > max) max = q; execbuf->flags = saved; } for (int i = 0; i < ARRAY_SIZE(ctx); i++) gem_context_destroy(fd, ctx[i]); igt_debug("Estimated qlen: {min:%d, max:%d}\n", min, max); return min; } static void single(int fd, uint32_t handle, const struct intel_execution_engine2 *e2, unsigned flags, const int ncpus, int timeout) { struct drm_i915_gem_execbuffer2 execbuf; struct drm_i915_gem_exec_object2 obj; struct drm_i915_gem_relocation_entry reloc; uint32_t contexts[64]; struct { double elapsed; unsigned long count; } *shared; int n; shared = mmap(NULL, 4096, PROT_WRITE, MAP_SHARED | MAP_ANON, -1, 0); igt_assert(shared != MAP_FAILED); for (n = 0; n < 64; n++) { if (flags & QUEUE) contexts[n] = gem_queue_create(fd); else contexts[n] = gem_context_create(fd); if (gem_context_has_engine_map(fd, 0)) gem_context_set_all_engines(fd, contexts[n]); } memset(&obj, 0, sizeof(obj)); obj.handle = handle; if (flags & INTERRUPTIBLE) { /* Be tricksy and force a relocation every batch so that * we don't emit the batch but just do MI_SET_CONTEXT */ memset(&reloc, 0, sizeof(reloc)); reloc.offset = 1024; reloc.read_domains = I915_GEM_DOMAIN_INSTRUCTION; obj.relocs_ptr = to_user_pointer(&reloc); obj.relocation_count = 1; } memset(&execbuf, 0, sizeof(execbuf)); execbuf.buffers_ptr = to_user_pointer(&obj); execbuf.buffer_count = 1; execbuf.rsvd1 = contexts[0]; execbuf.flags = e2->flags; execbuf.flags |= LOCAL_I915_EXEC_HANDLE_LUT; execbuf.flags |= LOCAL_I915_EXEC_NO_RELOC; igt_require(__gem_execbuf(fd, &execbuf) == 0); if (__gem_execbuf(fd, &execbuf)) { execbuf.flags = e2->flags; reloc.target_handle = obj.handle; gem_execbuf(fd, &execbuf); } gem_sync(fd, handle); igt_fork(child, ncpus) { struct timespec start, now; unsigned int count = 0; /* Warmup to bind all objects into each ctx before we begin */ for (int i = 0; i < ARRAY_SIZE(contexts); i++) { execbuf.rsvd1 = contexts[i]; gem_execbuf(fd, &execbuf); } gem_sync(fd, handle); clock_gettime(CLOCK_MONOTONIC, &start); do { igt_while_interruptible(flags & INTERRUPTIBLE) { for (int loop = 0; loop < 64; loop++) { execbuf.rsvd1 = contexts[loop % 64]; reloc.presumed_offset = -1; gem_execbuf(fd, &execbuf); } count += 64; } clock_gettime(CLOCK_MONOTONIC, &now); } while (elapsed(&start, &now) < timeout); gem_sync(fd, handle); clock_gettime(CLOCK_MONOTONIC, &now); igt_info("[%d] %s: %'u cycles: %.3fus%s\n", child, e2->name, count, elapsed(&start, &now) * 1e6 / count, flags & INTERRUPTIBLE ? " (interruptible)" : ""); shared[child].elapsed = elapsed(&start, &now); shared[child].count = count; } igt_waitchildren(); if (ncpus > 1) { unsigned long total = 0; double max = 0; for (n = 0; n < ncpus; n++) { total += shared[n].count; if (shared[n].elapsed > max) max = shared[n].elapsed; } igt_info("Total %s: %'lu cycles: %.3fus%s\n", e2->name, total, max*1e6 / total, flags & INTERRUPTIBLE ? " (interruptible)" : ""); } for (n = 0; n < 64; n++) gem_context_destroy(fd, contexts[n]); munmap(shared, 4096); } static void all(int fd, uint32_t handle, unsigned flags, int timeout) { struct drm_i915_gem_execbuffer2 execbuf; struct drm_i915_gem_exec_object2 obj[2]; struct intel_engine_data engines = { }; uint32_t contexts[65]; int n, qlen; engines = intel_init_engine_list(fd, 0); igt_require(engines.nengines); for (n = 0; n < ARRAY_SIZE(contexts); n++) { if (flags & QUEUE) contexts[n] = gem_queue_create(fd); else contexts[n] = gem_context_create(fd); gem_context_set_all_engines(fd, contexts[n]); } memset(obj, 0, sizeof(obj)); obj[1].handle = handle; memset(&execbuf, 0, sizeof(execbuf)); execbuf.buffers_ptr = to_user_pointer(obj + 1); execbuf.buffer_count = 1; execbuf.rsvd1 = contexts[0]; execbuf.flags |= LOCAL_I915_EXEC_HANDLE_LUT; execbuf.flags |= LOCAL_I915_EXEC_NO_RELOC; igt_require(__gem_execbuf(fd, &execbuf) == 0); gem_sync(fd, handle); qlen = measure_qlen(fd, &execbuf, &engines, timeout); igt_info("Using timing depth of %d batches\n", qlen); execbuf.buffers_ptr = to_user_pointer(obj); execbuf.buffer_count = 2; for (int pot = 2; pot <= 64; pot *= 2) { for (int nctx = pot - 1; nctx <= pot + 1; nctx++) { igt_fork(child, engines.nengines) { struct timespec start, now; unsigned int count = 0; obj[0].handle = gem_create(fd, 4096); execbuf.flags |= engines.engines[child].flags; for (int loop = 0; loop < ARRAY_SIZE(contexts); loop++) { execbuf.rsvd1 = contexts[loop]; gem_execbuf(fd, &execbuf); } gem_sync(fd, obj[0].handle); clock_gettime(CLOCK_MONOTONIC, &start); do { for (int loop = 0; loop < qlen; loop++) { execbuf.rsvd1 = contexts[loop % nctx]; gem_execbuf(fd, &execbuf); } count += qlen; gem_sync(fd, obj[0].handle); clock_gettime(CLOCK_MONOTONIC, &now); } while (elapsed(&start, &now) < timeout); gem_sync(fd, obj[0].handle); clock_gettime(CLOCK_MONOTONIC, &now); gem_close(fd, obj[0].handle); igt_info("[%d:%d] %s: %'u cycles: %.3fus%s (elapsed: %.3fs)\n", nctx, child, engines.engines[child].name, count, elapsed(&start, &now) * 1e6 / count, flags & INTERRUPTIBLE ? " (interruptible)" : "", elapsed(&start, &now)); } igt_waitchildren(); } } for (n = 0; n < ARRAY_SIZE(contexts); n++) gem_context_destroy(fd, contexts[n]); } igt_main { const int ncpus = sysconf(_SC_NPROCESSORS_ONLN); const struct intel_execution_engine2 *e2; const struct intel_execution_engine *e; static const struct { const char *name; unsigned int flags; bool (*require)(int fd); } phases[] = { { "", 0, NULL }, { "-interruptible", INTERRUPTIBLE, NULL }, { "-queue", QUEUE, gem_has_queues }, { "-queue-interruptible", QUEUE | INTERRUPTIBLE, gem_has_queues }, { } }; uint32_t light = 0, heavy; int fd = -1; igt_fixture { const uint32_t bbe = MI_BATCH_BUFFER_END; fd = drm_open_driver(DRIVER_INTEL); igt_require_gem(fd); gem_require_contexts(fd); light = gem_create(fd, 4096); gem_write(fd, light, 0, &bbe, sizeof(bbe)); heavy = gem_create(fd, 4096*1024); gem_write(fd, heavy, 4096*1024-sizeof(bbe), &bbe, sizeof(bbe)); igt_fork_hang_detector(fd); } /* Legacy testing must be first. */ for (e = intel_execution_engines; e->name; e++) { struct intel_execution_engine2 e2__; e2__ = gem_eb_flags_to_engine(e->exec_id | e->flags); if (e2__.flags == -1) continue; /* I915_EXEC_BSD with no ring selectors */ e2 = &e2__; for (typeof(*phases) *p = phases; p->name; p++) { igt_subtest_group { igt_fixture { gem_require_ring(fd, e2->flags); if (p->require) igt_require(p->require(fd)); } igt_subtest_f("legacy-%s%s", e->name, p->name) single(fd, light, e2, p->flags, 1, 5); igt_skip_on_simulation(); igt_subtest_f("legacy-%s-heavy%s", e->name, p->name) single(fd, heavy, e2, p->flags, 1, 5); igt_subtest_f("legacy-%s-forked%s", e->name, p->name) single(fd, light, e2, p->flags, ncpus, 150); igt_subtest_f("legacy-%s-forked-heavy%s", e->name, p->name) single(fd, heavy, e2, p->flags, ncpus, 150); } } } /* Must come after legacy subtests. */ __for_each_physical_engine(fd, e2) { for (typeof(*phases) *p = phases; p->name; p++) { igt_subtest_group { igt_fixture { if (p->require) igt_require(p->require(fd)); } igt_subtest_f("%s%s", e2->name, p->name) single(fd, light, e2, p->flags, 1, 5); igt_skip_on_simulation(); igt_subtest_f("%s-heavy%s", e2->name, p->name) single(fd, heavy, e2, p->flags, 1, 5); igt_subtest_f("%s-forked%s", e2->name, p->name) single(fd, light, e2, p->flags, ncpus, 150); igt_subtest_f("%s-forked-heavy%s", e2->name, p->name) single(fd, heavy, e2, p->flags, ncpus, 150); } } } igt_subtest("all-light") all(fd, light, 0, 5); igt_subtest("all-heavy") all(fd, heavy, 0, 5); igt_subtest_group { igt_fixture { igt_require(gem_has_queues(fd)); } igt_subtest("queue-light") all(fd, light, QUEUE, 5); igt_subtest("queue-heavy") all(fd, heavy, QUEUE, 5); } igt_fixture { igt_stop_hang_detector(); gem_close(fd, heavy); gem_close(fd, light); close(fd); } }