/* * Copyright © 2017 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #include "common/gen_decoder.h" #include "gen_disasm.h" #include void gen_batch_decode_ctx_init(struct gen_batch_decode_ctx *ctx, const struct gen_device_info *devinfo, FILE *fp, enum gen_batch_decode_flags flags, const char *xml_path, struct gen_batch_decode_bo (*get_bo)(void *, uint64_t), void *user_data) { memset(ctx, 0, sizeof(*ctx)); ctx->get_bo = get_bo; ctx->user_data = user_data; ctx->fp = fp; ctx->flags = flags; if (xml_path == NULL) ctx->spec = gen_spec_load(devinfo); else ctx->spec = gen_spec_load_from_path(devinfo, xml_path); ctx->disasm = gen_disasm_create(devinfo); } void gen_batch_decode_ctx_finish(struct gen_batch_decode_ctx *ctx) { gen_spec_destroy(ctx->spec); gen_disasm_destroy(ctx->disasm); } #define CSI "\e[" #define BLUE_HEADER CSI "0;44m" #define GREEN_HEADER CSI "1;42m" #define NORMAL CSI "0m" #define ARRAY_LENGTH(a) (sizeof (a) / sizeof (a)[0]) static void ctx_print_group(struct gen_batch_decode_ctx *ctx, struct gen_group *group, uint64_t address, const void *map) { gen_print_group(ctx->fp, group, address, map, 0, (ctx->flags & GEN_BATCH_DECODE_IN_COLOR) != 0); } static struct gen_batch_decode_bo ctx_get_bo(struct gen_batch_decode_ctx *ctx, uint64_t addr) { if (gen_spec_get_gen(ctx->spec) >= gen_make_gen(8,0)) { /* On Broadwell and above, we have 48-bit addresses which consume two * dwords. Some packets require that these get stored in a "canonical * form" which means that bit 47 is sign-extended through the upper * bits. In order to correctly handle those aub dumps, we need to mask * off the top 16 bits. */ addr &= (~0ull >> 16); } struct gen_batch_decode_bo bo = ctx->get_bo(ctx->user_data, addr); if (gen_spec_get_gen(ctx->spec) >= gen_make_gen(8,0)) bo.addr &= (~0ull >> 16); /* We may actually have an offset into the bo */ if (bo.map != NULL) { assert(bo.addr <= addr); uint64_t offset = addr - bo.addr; bo.map += offset; bo.addr += offset; bo.size -= offset; } return bo; } static void ctx_disassemble_program(struct gen_batch_decode_ctx *ctx, uint32_t ksp, const char *type) { if (!ctx->instruction_base.map) return; printf("\nReferenced %s:\n", type); gen_disasm_disassemble(ctx->disasm, (void *)ctx->instruction_base.map, ksp, ctx->fp); } /* Heuristic to determine whether a uint32_t is probably actually a float * (http://stackoverflow.com/a/2953466) */ static bool probably_float(uint32_t bits) { int exp = ((bits & 0x7f800000U) >> 23) - 127; uint32_t mant = bits & 0x007fffff; /* +- 0.0 */ if (exp == -127 && mant == 0) return true; /* +- 1 billionth to 1 billion */ if (-30 <= exp && exp <= 30) return true; /* some value with only a few binary digits */ if ((mant & 0x0000ffff) == 0) return true; return false; } static void ctx_print_buffer(struct gen_batch_decode_ctx *ctx, struct gen_batch_decode_bo bo, uint32_t read_length, uint32_t pitch) { const uint32_t *dw_end = bo.map + MIN2(bo.size, read_length); unsigned line_count = 0; for (const uint32_t *dw = bo.map; dw < dw_end; dw++) { if (line_count * 4 == pitch || line_count == 8) { fprintf(ctx->fp, "\n"); line_count = 0; } fprintf(ctx->fp, line_count == 0 ? " " : " "); if ((ctx->flags & GEN_BATCH_DECODE_FLOATS) && probably_float(*dw)) fprintf(ctx->fp, " %8.2f", *(float *) dw); else fprintf(ctx->fp, " 0x%08x", *dw); line_count++; } fprintf(ctx->fp, "\n"); } static void handle_state_base_address(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { struct gen_group *inst = gen_spec_find_instruction(ctx->spec, p); struct gen_field_iterator iter; gen_field_iterator_init(&iter, inst, p, 0, false); do { if (strcmp(iter.name, "Surface State Base Address") == 0) { ctx->surface_base = ctx_get_bo(ctx, iter.raw_value); } else if (strcmp(iter.name, "Dynamic State Base Address") == 0) { ctx->dynamic_base = ctx_get_bo(ctx, iter.raw_value); } else if (strcmp(iter.name, "Instruction Base Address") == 0) { ctx->instruction_base = ctx_get_bo(ctx, iter.raw_value); } } while (gen_field_iterator_next(&iter)); } static void dump_binding_table(struct gen_batch_decode_ctx *ctx, uint32_t offset, int count) { struct gen_group *strct = gen_spec_find_struct(ctx->spec, "RENDER_SURFACE_STATE"); if (strct == NULL) { fprintf(ctx->fp, "did not find RENDER_SURFACE_STATE info\n"); return; } /* If we don't know the actual count, guess. */ if (count < 0) count = 8; if (ctx->surface_base.map == NULL) { fprintf(ctx->fp, " binding table unavailable\n"); return; } if (offset % 32 != 0 || offset >= UINT16_MAX || offset >= ctx->surface_base.size) { fprintf(ctx->fp, " invalid binding table pointer\n"); return; } const uint32_t *pointers = ctx->surface_base.map + offset; for (int i = 0; i < count; i++) { if (pointers[i] == 0) continue; if (pointers[i] % 32 != 0 || (pointers[i] + strct->dw_length * 4) >= ctx->surface_base.size) { fprintf(ctx->fp, "pointer %u: %08x \n", i, pointers[i]); continue; } fprintf(ctx->fp, "pointer %u: %08x\n", i, pointers[i]); ctx_print_group(ctx, strct, ctx->surface_base.addr + pointers[i], ctx->surface_base.map + pointers[i]); } } static void dump_samplers(struct gen_batch_decode_ctx *ctx, uint32_t offset, int count) { struct gen_group *strct = gen_spec_find_struct(ctx->spec, "SAMPLER_STATE"); /* If we don't know the actual count, guess. */ if (count < 0) count = 4; if (ctx->dynamic_base.map == NULL) { fprintf(ctx->fp, " samplers unavailable\n"); return; } if (offset % 32 != 0 || offset >= ctx->dynamic_base.size) { fprintf(ctx->fp, " invalid sampler state pointer\n"); return; } uint64_t state_addr = ctx->dynamic_base.addr + offset; const void *state_map = ctx->dynamic_base.map + offset; for (int i = 0; i < count; i++) { fprintf(ctx->fp, "sampler state %d\n", i); ctx_print_group(ctx, strct, state_addr, state_map); state_addr += 16; state_map += 16; } } static void handle_media_interface_descriptor_load(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { if (ctx->dynamic_base.map == NULL) return; struct gen_group *inst = gen_spec_find_instruction(ctx->spec, p); struct gen_group *desc = gen_spec_find_struct(ctx->spec, "INTERFACE_DESCRIPTOR_DATA"); struct gen_field_iterator iter; gen_field_iterator_init(&iter, inst, p, 0, false); uint32_t descriptor_offset = 0; int descriptor_count = 0; do { if (strcmp(iter.name, "Interface Descriptor Data Start Address") == 0) { descriptor_offset = strtol(iter.value, NULL, 16); } else if (strcmp(iter.name, "Interface Descriptor Total Length") == 0) { descriptor_count = strtol(iter.value, NULL, 16) / (desc->dw_length * 4); } } while (gen_field_iterator_next(&iter)); uint64_t desc_addr = ctx->dynamic_base.addr + descriptor_offset; const uint32_t *desc_map = ctx->dynamic_base.map + descriptor_offset; for (int i = 0; i < descriptor_count; i++) { fprintf(ctx->fp, "descriptor %d: %08x\n", i, descriptor_offset); ctx_print_group(ctx, inst, desc_addr, desc_map); gen_field_iterator_init(&iter, desc, desc_map, 0, false); uint64_t ksp; uint32_t sampler_offset, sampler_count; uint32_t binding_table_offset, binding_entry_count; do { if (strcmp(iter.name, "Kernel Start Pointer") == 0) { ksp = strtoll(iter.value, NULL, 16); } else if (strcmp(iter.name, "Sampler State Pointer") == 0) { sampler_offset = strtol(iter.value, NULL, 16); } else if (strcmp(iter.name, "Sampler Count") == 0) { sampler_count = strtol(iter.value, NULL, 10); } else if (strcmp(iter.name, "Binding Table Pointer") == 0) { binding_table_offset = strtol(iter.value, NULL, 16); } else if (strcmp(iter.name, "Binding Table Entry Count") == 0) { binding_entry_count = strtol(iter.value, NULL, 10); } } while (gen_field_iterator_next(&iter)); ctx_disassemble_program(ctx, ksp, "compute shader"); printf("\n"); dump_samplers(ctx, sampler_offset, sampler_count); dump_binding_table(ctx, binding_table_offset, binding_entry_count); desc_map += desc->dw_length; desc_addr += desc->dw_length * 4; } } static void handle_3dstate_vertex_buffers(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { struct gen_group *inst = gen_spec_find_instruction(ctx->spec, p); struct gen_batch_decode_bo vb = {}; uint32_t vb_size = 0; int index = -1; int pitch = -1; bool ready = false; struct gen_field_iterator iter; gen_field_iterator_init(&iter, inst, p, 0, false); do { if (strcmp(iter.name, "Vertex Buffer Index") == 0) { index = iter.raw_value; } else if (strcmp(iter.name, "Buffer Pitch") == 0) { pitch = iter.raw_value; } else if (strcmp(iter.name, "Buffer Starting Address") == 0) { vb = ctx_get_bo(ctx, iter.raw_value); } else if (strcmp(iter.name, "Buffer Size") == 0) { vb_size = iter.raw_value; ready = true; } else if (strcmp(iter.name, "End Address") == 0) { if (vb.map && iter.raw_value >= vb.addr) vb_size = iter.raw_value - vb.addr; else vb_size = 0; ready = true; } if (!ready) continue; fprintf(ctx->fp, "vertex buffer %d, size %d\n", index, vb_size); if (vb.map == NULL) { fprintf(ctx->fp, " buffer contents unavailable\n"); continue; } if (vb.map == 0 || vb_size == 0) continue; ctx_print_buffer(ctx, vb, vb_size, pitch); vb.map = NULL; vb_size = 0; index = -1; pitch = -1; ready = false; } while (gen_field_iterator_next(&iter)); } static void handle_3dstate_index_buffer(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { struct gen_group *inst = gen_spec_find_instruction(ctx->spec, p); struct gen_batch_decode_bo ib = {}; uint32_t ib_size = 0; uint32_t format = 0; struct gen_field_iterator iter; gen_field_iterator_init(&iter, inst, p, 0, false); do { if (strcmp(iter.name, "Index Format") == 0) { format = iter.raw_value; } else if (strcmp(iter.name, "Buffer Starting Address") == 0) { ib = ctx_get_bo(ctx, iter.raw_value); } else if (strcmp(iter.name, "Buffer Size") == 0) { ib_size = iter.raw_value; } } while (gen_field_iterator_next(&iter)); if (ib.map == NULL) { fprintf(ctx->fp, " buffer contents unavailable\n"); return; } const void *m = ib.map; const void *ib_end = ib.map + MIN2(ib.size, ib_size); for (int i = 0; m < ib_end && i < 10; i++) { switch (format) { case 0: fprintf(ctx->fp, "%3d ", *(uint8_t *)m); m += 1; break; case 1: fprintf(ctx->fp, "%3d ", *(uint16_t *)m); m += 2; break; case 2: fprintf(ctx->fp, "%3d ", *(uint32_t *)m); m += 4; break; } } if (m < ib_end) fprintf(ctx->fp, "..."); fprintf(ctx->fp, "\n"); } static void decode_single_ksp(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { struct gen_group *inst = gen_spec_find_instruction(ctx->spec, p); uint64_t ksp = 0; bool is_simd8 = false; /* vertex shaders on Gen8+ only */ bool is_enabled = true; struct gen_field_iterator iter; gen_field_iterator_init(&iter, inst, p, 0, false); do { if (strcmp(iter.name, "Kernel Start Pointer") == 0) { ksp = iter.raw_value; } else if (strcmp(iter.name, "SIMD8 Dispatch Enable") == 0) { is_simd8 = iter.raw_value; } else if (strcmp(iter.name, "Dispatch Mode") == 0) { is_simd8 = strcmp(iter.value, "SIMD8") == 0; } else if (strcmp(iter.name, "Dispatch Enable") == 0) { is_simd8 = strcmp(iter.value, "SIMD8") == 0; } else if (strcmp(iter.name, "Enable") == 0) { is_enabled = iter.raw_value; } } while (gen_field_iterator_next(&iter)); const char *type = strcmp(inst->name, "VS_STATE") == 0 ? "vertex shader" : strcmp(inst->name, "GS_STATE") == 0 ? "geometry shader" : strcmp(inst->name, "SF_STATE") == 0 ? "strips and fans shader" : strcmp(inst->name, "CLIP_STATE") == 0 ? "clip shader" : strcmp(inst->name, "3DSTATE_DS") == 0 ? "tessellation control shader" : strcmp(inst->name, "3DSTATE_HS") == 0 ? "tessellation evaluation shader" : strcmp(inst->name, "3DSTATE_VS") == 0 ? (is_simd8 ? "SIMD8 vertex shader" : "vec4 vertex shader") : strcmp(inst->name, "3DSTATE_GS") == 0 ? (is_simd8 ? "SIMD8 geometry shader" : "vec4 geometry shader") : NULL; if (is_enabled) { ctx_disassemble_program(ctx, ksp, type); printf("\n"); } } static void decode_ps_kernels(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { struct gen_group *inst = gen_spec_find_instruction(ctx->spec, p); uint64_t ksp[3] = {0, 0, 0}; bool enabled[3] = {false, false, false}; struct gen_field_iterator iter; gen_field_iterator_init(&iter, inst, p, 0, false); do { if (strncmp(iter.name, "Kernel Start Pointer ", strlen("Kernel Start Pointer ")) == 0) { int idx = iter.name[strlen("Kernel Start Pointer ")] - '0'; ksp[idx] = strtol(iter.value, NULL, 16); } else if (strcmp(iter.name, "8 Pixel Dispatch Enable") == 0) { enabled[0] = strcmp(iter.value, "true") == 0; } else if (strcmp(iter.name, "16 Pixel Dispatch Enable") == 0) { enabled[1] = strcmp(iter.value, "true") == 0; } else if (strcmp(iter.name, "32 Pixel Dispatch Enable") == 0) { enabled[2] = strcmp(iter.value, "true") == 0; } } while (gen_field_iterator_next(&iter)); /* Reorder KSPs to be [8, 16, 32] instead of the hardware order. */ if (enabled[0] + enabled[1] + enabled[2] == 1) { if (enabled[1]) { ksp[1] = ksp[0]; ksp[0] = 0; } else if (enabled[2]) { ksp[2] = ksp[0]; ksp[0] = 0; } } else { uint64_t tmp = ksp[1]; ksp[1] = ksp[2]; ksp[2] = tmp; } if (enabled[0]) ctx_disassemble_program(ctx, ksp[0], "SIMD8 fragment shader"); if (enabled[1]) ctx_disassemble_program(ctx, ksp[1], "SIMD16 fragment shader"); if (enabled[2]) ctx_disassemble_program(ctx, ksp[2], "SIMD32 fragment shader"); fprintf(ctx->fp, "\n"); } static void decode_3dstate_constant(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { struct gen_group *inst = gen_spec_find_instruction(ctx->spec, p); uint32_t read_length[4]; struct gen_batch_decode_bo buffer[4]; memset(buffer, 0, sizeof(buffer)); int rlidx = 0, bidx = 0; struct gen_field_iterator iter; gen_field_iterator_init(&iter, inst, p, 0, false); do { if (strcmp(iter.name, "Read Length") == 0) { read_length[rlidx++] = iter.raw_value; } else if (strcmp(iter.name, "Buffer") == 0) { buffer[bidx++] = ctx_get_bo(ctx, iter.raw_value); } } while (gen_field_iterator_next(&iter)); for (int i = 0; i < 4; i++) { if (read_length[i] == 0 || buffer[i].map == NULL) continue; unsigned size = read_length[i] * 32; fprintf(ctx->fp, "constant buffer %d, size %u\n", i, size); ctx_print_buffer(ctx, buffer[i], size, 0); } } static void decode_3dstate_binding_table_pointers(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { dump_binding_table(ctx, p[1], -1); } static void decode_3dstate_sampler_state_pointers(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { dump_samplers(ctx, p[1], -1); } static void decode_3dstate_sampler_state_pointers_gen6(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { dump_samplers(ctx, p[1], -1); dump_samplers(ctx, p[2], -1); dump_samplers(ctx, p[3], -1); } static bool str_ends_with(const char *str, const char *end) { int offset = strlen(str) - strlen(end); if (offset < 0) return false; return strcmp(str + offset, end) == 0; } static void decode_dynamic_state_pointers(struct gen_batch_decode_ctx *ctx, const char *struct_type, const uint32_t *p, int count) { if (ctx->dynamic_base.map == NULL) { fprintf(ctx->fp, " dynamic %s state unavailable\n", struct_type); return; } struct gen_group *inst = gen_spec_find_instruction(ctx->spec, p); struct gen_group *state = gen_spec_find_struct(ctx->spec, struct_type); uint32_t state_offset; struct gen_field_iterator iter; gen_field_iterator_init(&iter, inst, p, 0, false); do { if (str_ends_with(iter.name, "Pointer")) { state_offset = iter.raw_value; break; } } while (gen_field_iterator_next(&iter)); uint32_t state_addr = ctx->dynamic_base.addr + state_offset; const uint32_t *state_map = ctx->dynamic_base.map + state_offset; for (int i = 0; i < count; i++) { fprintf(ctx->fp, "%s %d\n", struct_type, i); ctx_print_group(ctx, state, state_offset, state_map); state_addr += state->dw_length * 4; state_map += state->dw_length; } } static void decode_3dstate_viewport_state_pointers_cc(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { decode_dynamic_state_pointers(ctx, "CC_VIEWPORT", p, 4); } static void decode_3dstate_viewport_state_pointers_sf_clip(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { decode_dynamic_state_pointers(ctx, "SF_CLIP_VIEWPORT", p, 4); } static void decode_3dstate_blend_state_pointers(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { decode_dynamic_state_pointers(ctx, "BLEND_STATE", p, 1); } static void decode_3dstate_cc_state_pointers(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { decode_dynamic_state_pointers(ctx, "COLOR_CALC_STATE", p, 1); } static void decode_3dstate_scissor_state_pointers(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { decode_dynamic_state_pointers(ctx, "SCISSOR_RECT", p, 1); } static void decode_load_register_imm(struct gen_batch_decode_ctx *ctx, const uint32_t *p) { struct gen_group *reg = gen_spec_find_register(ctx->spec, p[1]); if (reg != NULL) { fprintf(ctx->fp, "register %s (0x%x): 0x%x\n", reg->name, reg->register_offset, p[2]); ctx_print_group(ctx, reg, reg->register_offset, &p[2]); } } struct custom_decoder { const char *cmd_name; void (*decode)(struct gen_batch_decode_ctx *ctx, const uint32_t *p); } custom_decoders[] = { { "STATE_BASE_ADDRESS", handle_state_base_address }, { "MEDIA_INTERFACE_DESCRIPTOR_LOAD", handle_media_interface_descriptor_load }, { "3DSTATE_VERTEX_BUFFERS", handle_3dstate_vertex_buffers }, { "3DSTATE_INDEX_BUFFER", handle_3dstate_index_buffer }, { "3DSTATE_VS", decode_single_ksp }, { "3DSTATE_GS", decode_single_ksp }, { "3DSTATE_DS", decode_single_ksp }, { "3DSTATE_HS", decode_single_ksp }, { "3DSTATE_PS", decode_ps_kernels }, { "3DSTATE_CONSTANT_VS", decode_3dstate_constant }, { "3DSTATE_CONSTANT_GS", decode_3dstate_constant }, { "3DSTATE_CONSTANT_PS", decode_3dstate_constant }, { "3DSTATE_CONSTANT_HS", decode_3dstate_constant }, { "3DSTATE_CONSTANT_DS", decode_3dstate_constant }, { "3DSTATE_BINDING_TABLE_POINTERS_VS", decode_3dstate_binding_table_pointers }, { "3DSTATE_BINDING_TABLE_POINTERS_HS", decode_3dstate_binding_table_pointers }, { "3DSTATE_BINDING_TABLE_POINTERS_DS", decode_3dstate_binding_table_pointers }, { "3DSTATE_BINDING_TABLE_POINTERS_GS", decode_3dstate_binding_table_pointers }, { "3DSTATE_BINDING_TABLE_POINTERS_PS", decode_3dstate_binding_table_pointers }, { "3DSTATE_SAMPLER_STATE_POINTERS_VS", decode_3dstate_sampler_state_pointers }, { "3DSTATE_SAMPLER_STATE_POINTERS_HS", decode_3dstate_sampler_state_pointers }, { "3DSTATE_SAMPLER_STATE_POINTERS_DS", decode_3dstate_sampler_state_pointers }, { "3DSTATE_SAMPLER_STATE_POINTERS_GS", decode_3dstate_sampler_state_pointers }, { "3DSTATE_SAMPLER_STATE_POINTERS_PS", decode_3dstate_sampler_state_pointers }, { "3DSTATE_SAMPLER_STATE_POINTERS", decode_3dstate_sampler_state_pointers_gen6 }, { "3DSTATE_VIEWPORT_STATE_POINTERS_CC", decode_3dstate_viewport_state_pointers_cc }, { "3DSTATE_VIEWPORT_STATE_POINTERS_SF_CLIP", decode_3dstate_viewport_state_pointers_sf_clip }, { "3DSTATE_BLEND_STATE_POINTERS", decode_3dstate_blend_state_pointers }, { "3DSTATE_CC_STATE_POINTERS", decode_3dstate_cc_state_pointers }, { "3DSTATE_SCISSOR_STATE_POINTERS", decode_3dstate_scissor_state_pointers }, { "MI_LOAD_REGISTER_IMM", decode_load_register_imm } }; static inline uint64_t get_address(struct gen_spec *spec, const uint32_t *p) { /* Addresses are always guaranteed to be page-aligned and sometimes * hardware packets have extra stuff stuffed in the bottom 12 bits. */ uint64_t addr = p[0] & ~0xfffu; if (gen_spec_get_gen(spec) >= gen_make_gen(8,0)) { /* On Broadwell and above, we have 48-bit addresses which consume two * dwords. Some packets require that these get stored in a "canonical * form" which means that bit 47 is sign-extended through the upper * bits. In order to correctly handle those aub dumps, we need to mask * off the top 16 bits. */ addr |= ((uint64_t)p[1] & 0xffff) << 32; } return addr; } void gen_print_batch(struct gen_batch_decode_ctx *ctx, const uint32_t *batch, uint32_t batch_size, uint64_t batch_addr) { const uint32_t *p, *end = batch + batch_size; int length; struct gen_group *inst; for (p = batch; p < end; p += length) { inst = gen_spec_find_instruction(ctx->spec, p); length = gen_group_get_length(inst, p); assert(inst == NULL || length > 0); length = MAX2(1, length); if (inst == NULL) { fprintf(ctx->fp, "unknown instruction %08x\n", p[0]); continue; } const char *color, *reset_color; uint64_t offset; const char *inst_name = gen_group_get_name(inst); if (ctx->flags & GEN_BATCH_DECODE_IN_COLOR) { reset_color = NORMAL; if (ctx->flags & GEN_BATCH_DECODE_FULL) { if (strcmp(inst_name, "MI_BATCH_BUFFER_START") == 0 || strcmp(inst_name, "MI_BATCH_BUFFER_END") == 0) color = GREEN_HEADER; else color = BLUE_HEADER; } else { color = NORMAL; } } else { color = ""; reset_color = ""; } if (ctx->flags & GEN_BATCH_DECODE_OFFSETS) offset = batch_addr + ((char *)p - (char *)batch); else offset = 0; fprintf(ctx->fp, "%s0x%08"PRIx64": 0x%08x: %-80s%s\n", color, offset, p[0], inst_name, reset_color); if (ctx->flags & GEN_BATCH_DECODE_FULL) { ctx_print_group(ctx, inst, offset, p); for (int i = 0; i < ARRAY_LENGTH(custom_decoders); i++) { if (strcmp(inst_name, custom_decoders[i].cmd_name) == 0) { custom_decoders[i].decode(ctx, p); break; } } } if (strcmp(inst_name, "MI_BATCH_BUFFER_START") == 0) { struct gen_batch_decode_bo next_batch; bool second_level; struct gen_field_iterator iter; gen_field_iterator_init(&iter, inst, p, 0, false); do { if (strcmp(iter.name, "Batch Buffer Start Address") == 0) { next_batch = ctx_get_bo(ctx, iter.raw_value); } else if (strcmp(iter.name, "Second Level Batch Buffer") == 0) { second_level = iter.raw_value; } } while (gen_field_iterator_next(&iter)); if (next_batch.map == NULL) { fprintf(ctx->fp, "Secondary batch at 0x%08"PRIx64" unavailable", next_batch.addr); } if (second_level) { /* MI_BATCH_BUFFER_START with "2nd Level Batch Buffer" set acts * like a subroutine call. Commands that come afterwards get * processed once the 2nd level batch buffer returns with * MI_BATCH_BUFFER_END. */ if (next_batch.map) { gen_print_batch(ctx, next_batch.map, next_batch.size, next_batch.addr); } } else { /* MI_BATCH_BUFFER_START with "2nd Level Batch Buffer" unset acts * like a goto. Nothing after it will ever get processed. In * order to prevent the recursion from growing, we just reset the * loop and continue; */ if (next_batch.map) { p = next_batch.map; end = next_batch.map + next_batch.size; length = 0; continue; } else { /* Nothing we can do */ break; } } } else if (strcmp(inst_name, "MI_BATCH_BUFFER_END") == 0) { break; } } }