/* * Copyright (C) 2017 Rob Clark * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * * Authors: * Rob Clark */ /* NOTE: see https://github.com/freedreno/freedreno/wiki/A5xx-Queries */ #include "freedreno_query_acc.h" #include "freedreno_resource.h" #include "fd5_context.h" #include "fd5_format.h" #include "fd5_query.h" struct PACKED fd5_query_sample { uint64_t start; uint64_t result; uint64_t stop; }; /* offset of a single field of an array of fd5_query_sample: */ #define query_sample_idx(aq, idx, field) \ fd_resource((aq)->prsc)->bo, \ (idx * sizeof(struct fd5_query_sample)) + \ offsetof(struct fd5_query_sample, field), \ 0, 0 /* offset of a single field of fd5_query_sample: */ #define query_sample(aq, field) \ query_sample_idx(aq, 0, field) /* * Occlusion Query: * * OCCLUSION_COUNTER and OCCLUSION_PREDICATE differ only in how they * interpret results */ static void occlusion_resume(struct fd_acc_query *aq, struct fd_batch *batch) { struct fd_ringbuffer *ring = batch->draw; OUT_PKT4(ring, REG_A5XX_RB_SAMPLE_COUNT_CONTROL, 1); OUT_RING(ring, A5XX_RB_SAMPLE_COUNT_CONTROL_COPY); OUT_PKT4(ring, REG_A5XX_RB_SAMPLE_COUNT_ADDR_LO, 2); OUT_RELOC(ring, query_sample(aq, start)); OUT_PKT7(ring, CP_EVENT_WRITE, 1); OUT_RING(ring, ZPASS_DONE); fd_reset_wfi(batch); fd5_context(batch->ctx)->samples_passed_queries++; } static void occlusion_pause(struct fd_acc_query *aq, struct fd_batch *batch) { struct fd_ringbuffer *ring = batch->draw; OUT_PKT7(ring, CP_MEM_WRITE, 4); OUT_RELOC(ring, query_sample(aq, stop)); OUT_RING(ring, 0xffffffff); OUT_RING(ring, 0xffffffff); OUT_PKT7(ring, CP_WAIT_MEM_WRITES, 0); OUT_PKT4(ring, REG_A5XX_RB_SAMPLE_COUNT_CONTROL, 1); OUT_RING(ring, A5XX_RB_SAMPLE_COUNT_CONTROL_COPY); OUT_PKT4(ring, REG_A5XX_RB_SAMPLE_COUNT_ADDR_LO, 2); OUT_RELOC(ring, query_sample(aq, stop)); OUT_PKT7(ring, CP_EVENT_WRITE, 1); OUT_RING(ring, ZPASS_DONE); fd_reset_wfi(batch); OUT_PKT7(ring, CP_WAIT_REG_MEM, 6); OUT_RING(ring, 0x00000014); // XXX OUT_RELOC(ring, query_sample(aq, stop)); OUT_RING(ring, 0xffffffff); OUT_RING(ring, 0xffffffff); OUT_RING(ring, 0x00000010); // XXX /* result += stop - start: */ OUT_PKT7(ring, CP_MEM_TO_MEM, 9); OUT_RING(ring, CP_MEM_TO_MEM_0_DOUBLE | CP_MEM_TO_MEM_0_NEG_C); OUT_RELOC(ring, query_sample(aq, result)); /* dst */ OUT_RELOC(ring, query_sample(aq, result)); /* srcA */ OUT_RELOC(ring, query_sample(aq, stop)); /* srcB */ OUT_RELOC(ring, query_sample(aq, start)); /* srcC */ fd5_context(batch->ctx)->samples_passed_queries--; } static void occlusion_counter_result(struct fd_acc_query *aq, void *buf, union pipe_query_result *result) { struct fd5_query_sample *sp = buf; result->u64 = sp->result; } static void occlusion_predicate_result(struct fd_acc_query *aq, void *buf, union pipe_query_result *result) { struct fd5_query_sample *sp = buf; result->b = !!sp->result; } static const struct fd_acc_sample_provider occlusion_counter = { .query_type = PIPE_QUERY_OCCLUSION_COUNTER, .size = sizeof(struct fd5_query_sample), .resume = occlusion_resume, .pause = occlusion_pause, .result = occlusion_counter_result, }; static const struct fd_acc_sample_provider occlusion_predicate = { .query_type = PIPE_QUERY_OCCLUSION_PREDICATE, .size = sizeof(struct fd5_query_sample), .resume = occlusion_resume, .pause = occlusion_pause, .result = occlusion_predicate_result, }; static const struct fd_acc_sample_provider occlusion_predicate_conservative = { .query_type = PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE, .size = sizeof(struct fd5_query_sample), .resume = occlusion_resume, .pause = occlusion_pause, .result = occlusion_predicate_result, }; /* * Timestamp Queries: */ static void timestamp_resume(struct fd_acc_query *aq, struct fd_batch *batch) { struct fd_ringbuffer *ring = batch->draw; OUT_PKT7(ring, CP_EVENT_WRITE, 4); OUT_RING(ring, CP_EVENT_WRITE_0_EVENT(RB_DONE_TS) | CP_EVENT_WRITE_0_TIMESTAMP); OUT_RELOC(ring, query_sample(aq, start)); OUT_RING(ring, 0x00000000); fd_reset_wfi(batch); } static void timestamp_pause(struct fd_acc_query *aq, struct fd_batch *batch) { struct fd_ringbuffer *ring = batch->draw; OUT_PKT7(ring, CP_EVENT_WRITE, 4); OUT_RING(ring, CP_EVENT_WRITE_0_EVENT(RB_DONE_TS) | CP_EVENT_WRITE_0_TIMESTAMP); OUT_RELOC(ring, query_sample(aq, stop)); OUT_RING(ring, 0x00000000); fd_reset_wfi(batch); fd_wfi(batch, ring); /* result += stop - start: */ OUT_PKT7(ring, CP_MEM_TO_MEM, 9); OUT_RING(ring, CP_MEM_TO_MEM_0_DOUBLE | CP_MEM_TO_MEM_0_NEG_C); OUT_RELOC(ring, query_sample(aq, result)); /* dst */ OUT_RELOC(ring, query_sample(aq, result)); /* srcA */ OUT_RELOC(ring, query_sample(aq, stop)); /* srcB */ OUT_RELOC(ring, query_sample(aq, start)); /* srcC */ } static uint64_t ticks_to_ns(uint32_t ts) { /* This is based on the 19.2MHz always-on rbbm timer. * * TODO we should probably query this value from kernel.. */ return ts * (1000000000 / 19200000); } static void time_elapsed_accumulate_result(struct fd_acc_query *aq, void *buf, union pipe_query_result *result) { struct fd5_query_sample *sp = buf; result->u64 = ticks_to_ns(sp->result); } static void timestamp_accumulate_result(struct fd_acc_query *aq, void *buf, union pipe_query_result *result) { struct fd5_query_sample *sp = buf; result->u64 = ticks_to_ns(sp->result); } static const struct fd_acc_sample_provider time_elapsed = { .query_type = PIPE_QUERY_TIME_ELAPSED, .always = true, .size = sizeof(struct fd5_query_sample), .resume = timestamp_resume, .pause = timestamp_pause, .result = time_elapsed_accumulate_result, }; /* NOTE: timestamp query isn't going to give terribly sensible results * on a tiler. But it is needed by qapitrace profile heatmap. If you * add in a binning pass, the results get even more non-sensical. So * we just return the timestamp on the first tile and hope that is * kind of good enough. */ static const struct fd_acc_sample_provider timestamp = { .query_type = PIPE_QUERY_TIMESTAMP, .always = true, .size = sizeof(struct fd5_query_sample), .resume = timestamp_resume, .pause = timestamp_pause, .result = timestamp_accumulate_result, }; /* * Performance Counter (batch) queries: * * Only one of these is active at a time, per design of the gallium * batch_query API design. On perfcntr query tracks N query_types, * each of which has a 'fd_batch_query_entry' that maps it back to * the associated group and counter. */ struct fd_batch_query_entry { uint8_t gid; /* group-id */ uint8_t cid; /* countable-id within the group */ }; struct fd_batch_query_data { struct fd_screen *screen; unsigned num_query_entries; struct fd_batch_query_entry query_entries[]; }; static void perfcntr_resume(struct fd_acc_query *aq, struct fd_batch *batch) { struct fd_batch_query_data *data = aq->query_data; struct fd_screen *screen = data->screen; struct fd_ringbuffer *ring = batch->draw; unsigned counters_per_group[screen->num_perfcntr_groups]; memset(counters_per_group, 0, sizeof(counters_per_group)); fd_wfi(batch, ring); /* configure performance counters for the requested queries: */ for (unsigned i = 0; i < data->num_query_entries; i++) { struct fd_batch_query_entry *entry = &data->query_entries[i]; const struct fd_perfcntr_group *g = &screen->perfcntr_groups[entry->gid]; unsigned counter_idx = counters_per_group[entry->gid]++; debug_assert(counter_idx < g->num_counters); OUT_PKT4(ring, g->counters[counter_idx].select_reg, 1); OUT_RING(ring, g->countables[entry->cid].selector); } memset(counters_per_group, 0, sizeof(counters_per_group)); /* and snapshot the start values */ for (unsigned i = 0; i < data->num_query_entries; i++) { struct fd_batch_query_entry *entry = &data->query_entries[i]; const struct fd_perfcntr_group *g = &screen->perfcntr_groups[entry->gid]; unsigned counter_idx = counters_per_group[entry->gid]++; const struct fd_perfcntr_counter *counter = &g->counters[counter_idx]; OUT_PKT7(ring, CP_REG_TO_MEM, 3); OUT_RING(ring, CP_REG_TO_MEM_0_64B | CP_REG_TO_MEM_0_REG(counter->counter_reg_lo)); OUT_RELOC(ring, query_sample_idx(aq, i, start)); } } static void perfcntr_pause(struct fd_acc_query *aq, struct fd_batch *batch) { struct fd_batch_query_data *data = aq->query_data; struct fd_screen *screen = data->screen; struct fd_ringbuffer *ring = batch->draw; unsigned counters_per_group[screen->num_perfcntr_groups]; memset(counters_per_group, 0, sizeof(counters_per_group)); fd_wfi(batch, ring); /* TODO do we need to bother to turn anything off? */ /* snapshot the end values: */ for (unsigned i = 0; i < data->num_query_entries; i++) { struct fd_batch_query_entry *entry = &data->query_entries[i]; const struct fd_perfcntr_group *g = &screen->perfcntr_groups[entry->gid]; unsigned counter_idx = counters_per_group[entry->gid]++; const struct fd_perfcntr_counter *counter = &g->counters[counter_idx]; OUT_PKT7(ring, CP_REG_TO_MEM, 3); OUT_RING(ring, CP_REG_TO_MEM_0_64B | CP_REG_TO_MEM_0_REG(counter->counter_reg_lo)); OUT_RELOC(ring, query_sample_idx(aq, i, stop)); } /* and compute the result: */ for (unsigned i = 0; i < data->num_query_entries; i++) { /* result += stop - start: */ OUT_PKT7(ring, CP_MEM_TO_MEM, 9); OUT_RING(ring, CP_MEM_TO_MEM_0_DOUBLE | CP_MEM_TO_MEM_0_NEG_C); OUT_RELOC(ring, query_sample_idx(aq, i, result)); /* dst */ OUT_RELOC(ring, query_sample_idx(aq, i, result)); /* srcA */ OUT_RELOC(ring, query_sample_idx(aq, i, stop)); /* srcB */ OUT_RELOC(ring, query_sample_idx(aq, i, start)); /* srcC */ } } static void perfcntr_accumulate_result(struct fd_acc_query *aq, void *buf, union pipe_query_result *result) { struct fd_batch_query_data *data = aq->query_data; struct fd5_query_sample *sp = buf; for (unsigned i = 0; i < data->num_query_entries; i++) { result->batch[i].u64 = sp[i].result; } } static const struct fd_acc_sample_provider perfcntr = { .query_type = FD_QUERY_FIRST_PERFCNTR, .always = true, .resume = perfcntr_resume, .pause = perfcntr_pause, .result = perfcntr_accumulate_result, }; static struct pipe_query * fd5_create_batch_query(struct pipe_context *pctx, unsigned num_queries, unsigned *query_types) { struct fd_context *ctx = fd_context(pctx); struct fd_screen *screen = ctx->screen; struct fd_query *q; struct fd_acc_query *aq; struct fd_batch_query_data *data; data = CALLOC_VARIANT_LENGTH_STRUCT(fd_batch_query_data, num_queries * sizeof(data->query_entries[0])); data->screen = screen; data->num_query_entries = num_queries; /* validate the requested query_types and ensure we don't try * to request more query_types of a given group than we have * counters: */ unsigned counters_per_group[screen->num_perfcntr_groups]; memset(counters_per_group, 0, sizeof(counters_per_group)); for (unsigned i = 0; i < num_queries; i++) { unsigned idx = query_types[i] - FD_QUERY_FIRST_PERFCNTR; /* verify valid query_type, ie. is it actually a perfcntr? */ if ((query_types[i] < FD_QUERY_FIRST_PERFCNTR) || (idx >= screen->num_perfcntr_queries)) { debug_printf("invalid batch query query_type: %u\n", query_types[i]); goto error; } struct fd_batch_query_entry *entry = &data->query_entries[i]; struct pipe_driver_query_info *pq = &screen->perfcntr_queries[idx]; entry->gid = pq->group_id; /* the perfcntr_queries[] table flattens all the countables * for each group in series, ie: * * (G0,C0), .., (G0,Cn), (G1,C0), .., (G1,Cm), ... * * So to find the countable index just step back through the * table to find the first entry with the same group-id. */ while (pq > screen->perfcntr_queries) { pq--; if (pq->group_id == entry->gid) entry->cid++; } if (counters_per_group[entry->gid] >= screen->perfcntr_groups[entry->gid].num_counters) { debug_printf("too many counters for group %u\n", entry->gid); goto error; } counters_per_group[entry->gid]++; } q = fd_acc_create_query2(ctx, 0, 0, &perfcntr); aq = fd_acc_query(q); /* sample buffer size is based on # of queries: */ aq->size = num_queries * sizeof(struct fd5_query_sample); aq->query_data = data; return (struct pipe_query *)q; error: free(data); return NULL; } void fd5_query_context_init(struct pipe_context *pctx) { struct fd_context *ctx = fd_context(pctx); ctx->create_query = fd_acc_create_query; ctx->query_set_stage = fd_acc_query_set_stage; pctx->create_batch_query = fd5_create_batch_query; fd_acc_query_register_provider(pctx, &occlusion_counter); fd_acc_query_register_provider(pctx, &occlusion_predicate); fd_acc_query_register_provider(pctx, &occlusion_predicate_conservative); fd_acc_query_register_provider(pctx, &time_elapsed); fd_acc_query_register_provider(pctx, ×tamp); }