• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /**********************************************************
2  * Copyright 2008-2015 VMware, Inc.  All rights reserved.
3  *
4  * Permission is hereby granted, free of charge, to any person
5  * obtaining a copy of this software and associated documentation
6  * files (the "Software"), to deal in the Software without
7  * restriction, including without limitation the rights to use, copy,
8  * modify, merge, publish, distribute, sublicense, and/or sell copies
9  * of the Software, and to permit persons to whom the Software is
10  * furnished to do so, subject to the following conditions:
11  *
12  * The above copyright notice and this permission notice shall be
13  * included in all copies or substantial portions of the Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
16  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
17  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
18  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
19  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
20  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
21  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
22  * SOFTWARE.
23  *
24  **********************************************************/
25 
26 #include "pipe/p_state.h"
27 #include "pipe/p_context.h"
28 
29 #include "util/u_bitmask.h"
30 #include "util/u_memory.h"
31 
32 #include "svga_cmd.h"
33 #include "svga_context.h"
34 #include "svga_screen.h"
35 #include "svga_resource_buffer.h"
36 #include "svga_winsys.h"
37 #include "svga_debug.h"
38 
39 
40 /* Fixme: want a public base class for all pipe structs, even if there
41  * isn't much in them.
42  */
43 struct pipe_query {
44    int dummy;
45 };
46 
47 struct svga_query {
48    struct pipe_query base;
49    unsigned type;                  /**< PIPE_QUERY_x or SVGA_QUERY_x */
50    SVGA3dQueryType svga_type;      /**< SVGA3D_QUERYTYPE_x or unused */
51 
52    unsigned id;                    /** Per-context query identifier */
53    boolean active;                 /** TRUE if query is active */
54 
55    struct pipe_fence_handle *fence;
56 
57    /** For PIPE_QUERY_OCCLUSION_COUNTER / SVGA3D_QUERYTYPE_OCCLUSION */
58 
59    /* For VGPU9 */
60    struct svga_winsys_buffer *hwbuf;
61    volatile SVGA3dQueryResult *queryResult;
62 
63    /** For VGPU10 */
64    struct svga_winsys_gb_query *gb_query;
65    SVGA3dDXQueryFlags flags;
66    unsigned offset;                /**< offset to the gb_query memory */
67    struct pipe_query *predicate;   /** The associated query that can be used for predicate */
68 
69    /** For non-GPU SVGA_QUERY_x queries */
70    uint64_t begin_count, end_count;
71 };
72 
73 
74 /** cast wrapper */
75 static inline struct svga_query *
svga_query(struct pipe_query * q)76 svga_query(struct pipe_query *q)
77 {
78    return (struct svga_query *)q;
79 }
80 
81 /**
82  * VGPU9
83  */
84 
85 static bool
86 svga_get_query_result(struct pipe_context *pipe,
87                       struct pipe_query *q,
88                       bool wait,
89                       union pipe_query_result *result);
90 
91 static enum pipe_error
define_query_vgpu9(struct svga_context * svga,struct svga_query * sq)92 define_query_vgpu9(struct svga_context *svga,
93                    struct svga_query *sq)
94 {
95    struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws;
96 
97    sq->hwbuf = svga_winsys_buffer_create(svga, 1,
98                                          SVGA_BUFFER_USAGE_PINNED,
99                                          sizeof *sq->queryResult);
100    if (!sq->hwbuf)
101       return PIPE_ERROR_OUT_OF_MEMORY;
102 
103    sq->queryResult = (SVGA3dQueryResult *)
104                      sws->buffer_map(sws, sq->hwbuf, PIPE_MAP_WRITE);
105    if (!sq->queryResult) {
106       sws->buffer_destroy(sws, sq->hwbuf);
107       return PIPE_ERROR_OUT_OF_MEMORY;
108    }
109 
110    sq->queryResult->totalSize = sizeof *sq->queryResult;
111    sq->queryResult->state = SVGA3D_QUERYSTATE_NEW;
112 
113    /* We request the buffer to be pinned and assume it is always mapped.
114     * The reason is that we don't want to wait for fences when checking the
115     * query status.
116     */
117    sws->buffer_unmap(sws, sq->hwbuf);
118 
119    return PIPE_OK;
120 }
121 
122 static void
begin_query_vgpu9(struct svga_context * svga,struct svga_query * sq)123 begin_query_vgpu9(struct svga_context *svga, struct svga_query *sq)
124 {
125    struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws;
126 
127    if (sq->queryResult->state == SVGA3D_QUERYSTATE_PENDING) {
128       /* The application doesn't care for the pending query result.
129        * We cannot let go of the existing buffer and just get a new one
130        * because its storage may be reused for other purposes and clobbered
131        * by the host when it determines the query result.  So the only
132        * option here is to wait for the existing query's result -- not a
133        * big deal, given that no sane application would do this.
134        */
135        uint64_t result;
136        svga_get_query_result(&svga->pipe, &sq->base, TRUE, (void*)&result);
137        assert(sq->queryResult->state != SVGA3D_QUERYSTATE_PENDING);
138    }
139 
140    sq->queryResult->state = SVGA3D_QUERYSTATE_NEW;
141    sws->fence_reference(sws, &sq->fence, NULL);
142 
143    SVGA_RETRY(svga, SVGA3D_BeginQuery(svga->swc, sq->svga_type));
144 }
145 
146 static void
end_query_vgpu9(struct svga_context * svga,struct svga_query * sq)147 end_query_vgpu9(struct svga_context *svga, struct svga_query *sq)
148 {
149    /* Set to PENDING before sending EndQuery. */
150    sq->queryResult->state = SVGA3D_QUERYSTATE_PENDING;
151 
152    SVGA_RETRY(svga, SVGA3D_EndQuery(svga->swc, sq->svga_type, sq->hwbuf));
153 }
154 
155 static bool
get_query_result_vgpu9(struct svga_context * svga,struct svga_query * sq,bool wait,uint64_t * result)156 get_query_result_vgpu9(struct svga_context *svga, struct svga_query *sq,
157                        bool wait, uint64_t *result)
158 {
159    struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws;
160    SVGA3dQueryState state;
161 
162    if (!sq->fence) {
163       /* The query status won't be updated by the host unless
164        * SVGA_3D_CMD_WAIT_FOR_QUERY is emitted. Unfortunately this will cause
165        * a synchronous wait on the host.
166        */
167       SVGA_RETRY(svga, SVGA3D_WaitForQuery(svga->swc, sq->svga_type,
168                                            sq->hwbuf));
169       svga_context_flush(svga, &sq->fence);
170       assert(sq->fence);
171    }
172 
173    state = sq->queryResult->state;
174    if (state == SVGA3D_QUERYSTATE_PENDING) {
175       if (!wait)
176          return false;
177       sws->fence_finish(sws, sq->fence, PIPE_TIMEOUT_INFINITE,
178                         SVGA_FENCE_FLAG_QUERY);
179       state = sq->queryResult->state;
180    }
181 
182    assert(state == SVGA3D_QUERYSTATE_SUCCEEDED ||
183           state == SVGA3D_QUERYSTATE_FAILED);
184 
185    *result = (uint64_t)sq->queryResult->result32;
186    return true;
187 }
188 
189 
190 /**
191  * VGPU10
192  *
193  * There is one query mob allocated for each context to be shared by all
194  * query types. The mob is used to hold queries's state and result. Since
195  * each query result type is of different length, to ease the query allocation
196  * management, the mob is divided into memory blocks. Each memory block
197  * will hold queries of the same type. Multiple memory blocks can be allocated
198  * for a particular query type.
199  *
200  * Currently each memory block is of 184 bytes. We support up to 512
201  * memory blocks. The query memory size is arbitrary right now.
202  * Each occlusion query takes about 8 bytes. One memory block can accomodate
203  * 23 occlusion queries. 512 of those blocks can support up to 11K occlusion
204  * queries. That seems reasonable for now. If we think this limit is
205  * not enough, we can increase the limit or try to grow the mob in runtime.
206  * Note, SVGA device does not impose one mob per context for queries,
207  * we could allocate multiple mobs for queries; however, wddm KMD does not
208  * currently support that.
209  *
210  * Also note that the GL guest driver does not issue any of the
211  * following commands: DXMoveQuery, DXBindAllQuery & DXReadbackAllQuery.
212  */
213 #define SVGA_QUERY_MEM_BLOCK_SIZE    (sizeof(SVGADXQueryResultUnion) * 2)
214 #define SVGA_QUERY_MEM_SIZE          (512 * SVGA_QUERY_MEM_BLOCK_SIZE)
215 
216 struct svga_qmem_alloc_entry
217 {
218    unsigned start_offset;               /* start offset of the memory block */
219    unsigned block_index;                /* block index of the memory block */
220    unsigned query_size;                 /* query size in this memory block */
221    unsigned nquery;                     /* number of queries allocated */
222    struct util_bitmask *alloc_mask;     /* allocation mask */
223    struct svga_qmem_alloc_entry *next;  /* next memory block */
224 };
225 
226 
227 /**
228  * Allocate a memory block from the query object memory
229  * \return NULL if out of memory, else pointer to the query memory block
230  */
231 static struct svga_qmem_alloc_entry *
allocate_query_block(struct svga_context * svga)232 allocate_query_block(struct svga_context *svga)
233 {
234    int index;
235    unsigned offset;
236    struct svga_qmem_alloc_entry *alloc_entry = NULL;
237 
238    /* Find the next available query block */
239    index = util_bitmask_add(svga->gb_query_alloc_mask);
240 
241    if (index == UTIL_BITMASK_INVALID_INDEX)
242       return NULL;
243 
244    offset = index * SVGA_QUERY_MEM_BLOCK_SIZE;
245    if (offset >= svga->gb_query_len) {
246       unsigned i;
247 
248       /* Deallocate the out-of-range index */
249       util_bitmask_clear(svga->gb_query_alloc_mask, index);
250       index = -1;
251 
252       /**
253        * All the memory blocks are allocated, lets see if there is
254        * any empty memory block around that can be freed up.
255        */
256       for (i = 0; i < SVGA3D_QUERYTYPE_MAX && index == -1; i++) {
257          struct svga_qmem_alloc_entry *prev_alloc_entry = NULL;
258 
259          alloc_entry = svga->gb_query_map[i];
260          while (alloc_entry && index == -1) {
261             if (alloc_entry->nquery == 0) {
262                /* This memory block is empty, it can be recycled. */
263                if (prev_alloc_entry) {
264                   prev_alloc_entry->next = alloc_entry->next;
265                } else {
266                   svga->gb_query_map[i] = alloc_entry->next;
267                }
268                index = alloc_entry->block_index;
269             } else {
270                prev_alloc_entry = alloc_entry;
271                alloc_entry = alloc_entry->next;
272             }
273          }
274       }
275 
276       if (index == -1) {
277          debug_printf("Query memory object is full\n");
278          return NULL;
279       }
280    }
281 
282    if (!alloc_entry) {
283       assert(index != -1);
284       alloc_entry = CALLOC_STRUCT(svga_qmem_alloc_entry);
285       alloc_entry->block_index = index;
286    }
287 
288    return alloc_entry;
289 }
290 
291 /**
292  * Allocate a slot in the specified memory block.
293  * All slots in this memory block are of the same size.
294  *
295  * \return -1 if out of memory, else index of the query slot
296  */
297 static int
allocate_query_slot(struct svga_context * svga,struct svga_qmem_alloc_entry * alloc)298 allocate_query_slot(struct svga_context *svga,
299                     struct svga_qmem_alloc_entry *alloc)
300 {
301    int index;
302    unsigned offset;
303 
304    /* Find the next available slot */
305    index = util_bitmask_add(alloc->alloc_mask);
306 
307    if (index == UTIL_BITMASK_INVALID_INDEX)
308       return -1;
309 
310    offset = index * alloc->query_size;
311    if (offset >= SVGA_QUERY_MEM_BLOCK_SIZE)
312       return -1;
313 
314    alloc->nquery++;
315 
316    return index;
317 }
318 
319 /**
320  * Deallocate the specified slot in the memory block.
321  * If all slots are freed up, then deallocate the memory block
322  * as well, so it can be allocated for other query type
323  */
324 static void
deallocate_query_slot(struct svga_context * svga,struct svga_qmem_alloc_entry * alloc,unsigned index)325 deallocate_query_slot(struct svga_context *svga,
326                       struct svga_qmem_alloc_entry *alloc,
327                       unsigned index)
328 {
329    assert(index != UTIL_BITMASK_INVALID_INDEX);
330 
331    util_bitmask_clear(alloc->alloc_mask, index);
332    alloc->nquery--;
333 
334    /**
335     * Don't worry about deallocating the empty memory block here.
336     * The empty memory block will be recycled when no more memory block
337     * can be allocated.
338     */
339 }
340 
341 static struct svga_qmem_alloc_entry *
allocate_query_block_entry(struct svga_context * svga,unsigned len)342 allocate_query_block_entry(struct svga_context *svga,
343                            unsigned len)
344 {
345    struct svga_qmem_alloc_entry *alloc_entry;
346 
347    alloc_entry = allocate_query_block(svga);
348    if (!alloc_entry)
349       return NULL;
350 
351    assert(alloc_entry->block_index != -1);
352    alloc_entry->start_offset =
353       alloc_entry->block_index * SVGA_QUERY_MEM_BLOCK_SIZE;
354    alloc_entry->nquery = 0;
355    alloc_entry->alloc_mask = util_bitmask_create();
356    alloc_entry->next = NULL;
357    alloc_entry->query_size = len;
358 
359    return alloc_entry;
360 }
361 
362 /**
363  * Allocate a memory slot for a query of the specified type.
364  * It will first search through the memory blocks that are allocated
365  * for the query type. If no memory slot is available, it will try
366  * to allocate another memory block within the query object memory for
367  * this query type.
368  */
369 static int
allocate_query(struct svga_context * svga,SVGA3dQueryType type,unsigned len)370 allocate_query(struct svga_context *svga,
371                SVGA3dQueryType type,
372                unsigned len)
373 {
374    struct svga_qmem_alloc_entry *alloc_entry;
375    int slot_index = -1;
376    unsigned offset;
377 
378    assert(type < SVGA3D_QUERYTYPE_MAX);
379 
380    alloc_entry = svga->gb_query_map[type];
381 
382    if (!alloc_entry) {
383       /**
384        * No query memory block has been allocated for this query type,
385        * allocate one now
386        */
387       alloc_entry = allocate_query_block_entry(svga, len);
388       if (!alloc_entry)
389          return -1;
390       svga->gb_query_map[type] = alloc_entry;
391    }
392 
393    /* Allocate a slot within the memory block allocated for this query type */
394    slot_index = allocate_query_slot(svga, alloc_entry);
395 
396    if (slot_index == -1) {
397       /* This query memory block is full, allocate another one */
398       alloc_entry = allocate_query_block_entry(svga, len);
399       if (!alloc_entry)
400          return -1;
401       alloc_entry->next = svga->gb_query_map[type];
402       svga->gb_query_map[type] = alloc_entry;
403       slot_index = allocate_query_slot(svga, alloc_entry);
404    }
405 
406    assert(slot_index != -1);
407    offset = slot_index * len + alloc_entry->start_offset;
408 
409    return offset;
410 }
411 
412 
413 /**
414  * Deallocate memory slot allocated for the specified query
415  */
416 static void
deallocate_query(struct svga_context * svga,struct svga_query * sq)417 deallocate_query(struct svga_context *svga,
418                  struct svga_query *sq)
419 {
420    struct svga_qmem_alloc_entry *alloc_entry;
421    unsigned slot_index;
422    unsigned offset = sq->offset;
423 
424    alloc_entry = svga->gb_query_map[sq->svga_type];
425 
426    while (alloc_entry) {
427       if (offset >= alloc_entry->start_offset &&
428           offset < alloc_entry->start_offset + SVGA_QUERY_MEM_BLOCK_SIZE) {
429 
430          /* The slot belongs to this memory block, deallocate it */
431          slot_index = (offset - alloc_entry->start_offset) /
432                       alloc_entry->query_size;
433          deallocate_query_slot(svga, alloc_entry, slot_index);
434          alloc_entry = NULL;
435       } else {
436          alloc_entry = alloc_entry->next;
437       }
438    }
439 }
440 
441 
442 /**
443  * Destroy the gb query object and all the related query structures
444  */
445 static void
destroy_gb_query_obj(struct svga_context * svga)446 destroy_gb_query_obj(struct svga_context *svga)
447 {
448    struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws;
449    unsigned i;
450 
451    for (i = 0; i < SVGA3D_QUERYTYPE_MAX; i++) {
452       struct svga_qmem_alloc_entry *alloc_entry, *next;
453       alloc_entry = svga->gb_query_map[i];
454       while (alloc_entry) {
455          next = alloc_entry->next;
456          util_bitmask_destroy(alloc_entry->alloc_mask);
457          FREE(alloc_entry);
458          alloc_entry = next;
459       }
460       svga->gb_query_map[i] = NULL;
461    }
462 
463    if (svga->gb_query)
464       sws->query_destroy(sws, svga->gb_query);
465    svga->gb_query = NULL;
466 
467    util_bitmask_destroy(svga->gb_query_alloc_mask);
468 }
469 
470 /**
471  * Define query and create the gb query object if it is not already created.
472  * There is only one gb query object per context which will be shared by
473  * queries of all types.
474  */
475 static enum pipe_error
define_query_vgpu10(struct svga_context * svga,struct svga_query * sq,int resultLen)476 define_query_vgpu10(struct svga_context *svga,
477                     struct svga_query *sq, int resultLen)
478 {
479    struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws;
480    int qlen;
481    enum pipe_error ret = PIPE_OK;
482 
483    SVGA_DBG(DEBUG_QUERY, "%s\n", __FUNCTION__);
484 
485    if (svga->gb_query == NULL) {
486       /* Create a gb query object */
487       svga->gb_query = sws->query_create(sws, SVGA_QUERY_MEM_SIZE);
488       if (!svga->gb_query)
489          return PIPE_ERROR_OUT_OF_MEMORY;
490       svga->gb_query_len = SVGA_QUERY_MEM_SIZE;
491       memset (svga->gb_query_map, 0, sizeof(svga->gb_query_map));
492       svga->gb_query_alloc_mask = util_bitmask_create();
493 
494       /* Bind the query object to the context */
495       SVGA_RETRY(svga, svga->swc->query_bind(svga->swc, svga->gb_query,
496                                              SVGA_QUERY_FLAG_SET));
497    }
498 
499    sq->gb_query = svga->gb_query;
500 
501    /* Make sure query length is in multiples of 8 bytes */
502    qlen = align(resultLen + sizeof(SVGA3dQueryState), 8);
503 
504    /* Find a slot for this query in the gb object */
505    sq->offset = allocate_query(svga, sq->svga_type, qlen);
506    if (sq->offset == -1)
507       return PIPE_ERROR_OUT_OF_MEMORY;
508 
509    assert((sq->offset & 7) == 0);
510 
511    SVGA_DBG(DEBUG_QUERY, "   query type=%d qid=0x%x offset=%d\n",
512             sq->svga_type, sq->id, sq->offset);
513 
514    /**
515     * Send SVGA3D commands to define the query
516     */
517    SVGA_RETRY_OOM(svga, ret, SVGA3D_vgpu10_DefineQuery(svga->swc, sq->id,
518                                                        sq->svga_type,
519                                                        sq->flags));
520    if (ret != PIPE_OK)
521       return PIPE_ERROR_OUT_OF_MEMORY;
522 
523    SVGA_RETRY(svga, SVGA3D_vgpu10_BindQuery(svga->swc, sq->gb_query, sq->id));
524    SVGA_RETRY(svga, SVGA3D_vgpu10_SetQueryOffset(svga->swc, sq->id,
525                                                  sq->offset));
526 
527    return PIPE_OK;
528 }
529 
530 static void
destroy_query_vgpu10(struct svga_context * svga,struct svga_query * sq)531 destroy_query_vgpu10(struct svga_context *svga, struct svga_query *sq)
532 {
533    SVGA_RETRY(svga, SVGA3D_vgpu10_DestroyQuery(svga->swc, sq->id));
534 
535    /* Deallocate the memory slot allocated for this query */
536    deallocate_query(svga, sq);
537 }
538 
539 
540 /**
541  * Rebind queryies to the context.
542  */
543 static void
rebind_vgpu10_query(struct svga_context * svga)544 rebind_vgpu10_query(struct svga_context *svga)
545 {
546    SVGA_RETRY(svga, svga->swc->query_bind(svga->swc, svga->gb_query,
547                                           SVGA_QUERY_FLAG_REF));
548    svga->rebind.flags.query = FALSE;
549 }
550 
551 
552 static enum pipe_error
begin_query_vgpu10(struct svga_context * svga,struct svga_query * sq)553 begin_query_vgpu10(struct svga_context *svga, struct svga_query *sq)
554 {
555    struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws;
556    int status = 0;
557 
558    sws->fence_reference(sws, &sq->fence, NULL);
559 
560    /* Initialize the query state to NEW */
561    status = sws->query_init(sws, sq->gb_query, sq->offset, SVGA3D_QUERYSTATE_NEW);
562    if (status)
563       return PIPE_ERROR;
564 
565    if (svga->rebind.flags.query) {
566       rebind_vgpu10_query(svga);
567    }
568 
569    /* Send the BeginQuery command to the device */
570    SVGA_RETRY(svga, SVGA3D_vgpu10_BeginQuery(svga->swc, sq->id));
571    return PIPE_OK;
572 }
573 
574 static void
end_query_vgpu10(struct svga_context * svga,struct svga_query * sq)575 end_query_vgpu10(struct svga_context *svga, struct svga_query *sq)
576 {
577    if (svga->rebind.flags.query) {
578       rebind_vgpu10_query(svga);
579    }
580 
581    SVGA_RETRY(svga, SVGA3D_vgpu10_EndQuery(svga->swc, sq->id));
582 }
583 
584 static bool
get_query_result_vgpu10(struct svga_context * svga,struct svga_query * sq,bool wait,void * result,int resultLen)585 get_query_result_vgpu10(struct svga_context *svga, struct svga_query *sq,
586                         bool wait, void *result, int resultLen)
587 {
588    struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws;
589    SVGA3dQueryState queryState;
590 
591    if (svga->rebind.flags.query) {
592       rebind_vgpu10_query(svga);
593    }
594 
595    sws->query_get_result(sws, sq->gb_query, sq->offset, &queryState, result, resultLen);
596 
597    if (queryState != SVGA3D_QUERYSTATE_SUCCEEDED && !sq->fence) {
598       /* We don't have the query result yet, and the query hasn't been
599        * submitted.  We need to submit it now since the GL spec says
600        * "Querying the state for a given occlusion query forces that
601        * occlusion query to complete within a finite amount of time."
602        */
603       svga_context_flush(svga, &sq->fence);
604    }
605 
606    if (queryState == SVGA3D_QUERYSTATE_PENDING ||
607        queryState == SVGA3D_QUERYSTATE_NEW) {
608       if (!wait)
609          return false;
610       sws->fence_finish(sws, sq->fence, PIPE_TIMEOUT_INFINITE,
611                         SVGA_FENCE_FLAG_QUERY);
612       sws->query_get_result(sws, sq->gb_query, sq->offset, &queryState, result, resultLen);
613    }
614 
615    assert(queryState == SVGA3D_QUERYSTATE_SUCCEEDED ||
616           queryState == SVGA3D_QUERYSTATE_FAILED);
617 
618    return true;
619 }
620 
621 static struct pipe_query *
svga_create_query(struct pipe_context * pipe,unsigned query_type,unsigned index)622 svga_create_query(struct pipe_context *pipe,
623                   unsigned query_type,
624                   unsigned index)
625 {
626    struct svga_context *svga = svga_context(pipe);
627    struct svga_query *sq;
628    enum pipe_error ret;
629 
630    assert(query_type < SVGA_QUERY_MAX);
631 
632    sq = CALLOC_STRUCT(svga_query);
633    if (!sq)
634       goto fail;
635 
636    /* Allocate an integer ID for the query */
637    sq->id = util_bitmask_add(svga->query_id_bm);
638    if (sq->id == UTIL_BITMASK_INVALID_INDEX)
639       goto fail;
640 
641    SVGA_DBG(DEBUG_QUERY, "%s type=%d sq=0x%x id=%d\n", __FUNCTION__,
642             query_type, sq, sq->id);
643 
644    switch (query_type) {
645    case PIPE_QUERY_OCCLUSION_COUNTER:
646       sq->svga_type = SVGA3D_QUERYTYPE_OCCLUSION;
647       if (svga_have_vgpu10(svga)) {
648          ret = define_query_vgpu10(svga, sq,
649                                    sizeof(SVGADXOcclusionQueryResult));
650          if (ret != PIPE_OK)
651             goto fail;
652 
653          /**
654           * In OpenGL, occlusion counter query can be used in conditional
655           * rendering; however, in DX10, only OCCLUSION_PREDICATE query can
656           * be used for predication. Hence, we need to create an occlusion
657           * predicate query along with the occlusion counter query. So when
658           * the occlusion counter query is used for predication, the associated
659           * query of occlusion predicate type will be used
660           * in the SetPredication command.
661           */
662          sq->predicate = svga_create_query(pipe, PIPE_QUERY_OCCLUSION_PREDICATE, index);
663 
664       } else {
665          ret = define_query_vgpu9(svga, sq);
666          if (ret != PIPE_OK)
667             goto fail;
668       }
669       break;
670    case PIPE_QUERY_OCCLUSION_PREDICATE:
671    case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
672       if (svga_have_vgpu10(svga)) {
673          sq->svga_type = SVGA3D_QUERYTYPE_OCCLUSIONPREDICATE;
674          ret = define_query_vgpu10(svga, sq,
675                                    sizeof(SVGADXOcclusionPredicateQueryResult));
676          if (ret != PIPE_OK)
677             goto fail;
678       } else {
679          sq->svga_type = SVGA3D_QUERYTYPE_OCCLUSION;
680          ret = define_query_vgpu9(svga, sq);
681          if (ret != PIPE_OK)
682             goto fail;
683       }
684       break;
685    case PIPE_QUERY_PRIMITIVES_GENERATED:
686    case PIPE_QUERY_PRIMITIVES_EMITTED:
687    case PIPE_QUERY_SO_STATISTICS:
688       assert(svga_have_vgpu10(svga));
689 
690       /* Until the device supports the new query type for multiple streams,
691        * we will use the single stream query type for stream 0.
692        */
693       if (svga_have_sm5(svga) && index > 0) {
694          assert(index < 4);
695 
696          sq->svga_type = SVGA3D_QUERYTYPE_SOSTATS_STREAM0 + index;
697       }
698       else {
699          assert(index == 0);
700          sq->svga_type = SVGA3D_QUERYTYPE_STREAMOUTPUTSTATS;
701       }
702       ret = define_query_vgpu10(svga, sq,
703                                 sizeof(SVGADXStreamOutStatisticsQueryResult));
704       if (ret != PIPE_OK)
705          goto fail;
706       break;
707    case PIPE_QUERY_TIMESTAMP:
708       assert(svga_have_vgpu10(svga));
709       sq->svga_type = SVGA3D_QUERYTYPE_TIMESTAMP;
710       ret = define_query_vgpu10(svga, sq,
711                                 sizeof(SVGADXTimestampQueryResult));
712       if (ret != PIPE_OK)
713          goto fail;
714       break;
715    case SVGA_QUERY_NUM_DRAW_CALLS:
716    case SVGA_QUERY_NUM_FALLBACKS:
717    case SVGA_QUERY_NUM_FLUSHES:
718    case SVGA_QUERY_NUM_VALIDATIONS:
719    case SVGA_QUERY_NUM_BUFFERS_MAPPED:
720    case SVGA_QUERY_NUM_TEXTURES_MAPPED:
721    case SVGA_QUERY_NUM_BYTES_UPLOADED:
722    case SVGA_QUERY_NUM_COMMAND_BUFFERS:
723    case SVGA_QUERY_COMMAND_BUFFER_SIZE:
724    case SVGA_QUERY_SURFACE_WRITE_FLUSHES:
725    case SVGA_QUERY_MEMORY_USED:
726    case SVGA_QUERY_NUM_SHADERS:
727    case SVGA_QUERY_NUM_RESOURCES:
728    case SVGA_QUERY_NUM_STATE_OBJECTS:
729    case SVGA_QUERY_NUM_SURFACE_VIEWS:
730    case SVGA_QUERY_NUM_GENERATE_MIPMAP:
731    case SVGA_QUERY_NUM_READBACKS:
732    case SVGA_QUERY_NUM_RESOURCE_UPDATES:
733    case SVGA_QUERY_NUM_BUFFER_UPLOADS:
734    case SVGA_QUERY_NUM_CONST_BUF_UPDATES:
735    case SVGA_QUERY_NUM_CONST_UPDATES:
736    case SVGA_QUERY_NUM_FAILED_ALLOCATIONS:
737    case SVGA_QUERY_NUM_COMMANDS_PER_DRAW:
738    case SVGA_QUERY_NUM_SHADER_RELOCATIONS:
739    case SVGA_QUERY_NUM_SURFACE_RELOCATIONS:
740    case SVGA_QUERY_SHADER_MEM_USED:
741       break;
742    case SVGA_QUERY_FLUSH_TIME:
743    case SVGA_QUERY_MAP_BUFFER_TIME:
744       /* These queries need os_time_get() */
745       svga->hud.uses_time = TRUE;
746       break;
747 
748    default:
749       assert(!"unexpected query type in svga_create_query()");
750    }
751 
752    sq->type = query_type;
753 
754    return &sq->base;
755 
756 fail:
757    FREE(sq);
758    return NULL;
759 }
760 
761 static void
svga_destroy_query(struct pipe_context * pipe,struct pipe_query * q)762 svga_destroy_query(struct pipe_context *pipe, struct pipe_query *q)
763 {
764    struct svga_context *svga = svga_context(pipe);
765    struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws;
766    struct svga_query *sq;
767 
768    if (!q) {
769       destroy_gb_query_obj(svga);
770       return;
771    }
772 
773    sq = svga_query(q);
774 
775    SVGA_DBG(DEBUG_QUERY, "%s sq=0x%x id=%d\n", __FUNCTION__,
776             sq, sq->id);
777 
778    switch (sq->type) {
779    case PIPE_QUERY_OCCLUSION_COUNTER:
780    case PIPE_QUERY_OCCLUSION_PREDICATE:
781    case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
782       if (svga_have_vgpu10(svga)) {
783          /* make sure to also destroy any associated predicate query */
784          if (sq->predicate)
785             svga_destroy_query(pipe, sq->predicate);
786          destroy_query_vgpu10(svga, sq);
787       } else {
788          sws->buffer_destroy(sws, sq->hwbuf);
789       }
790       sws->fence_reference(sws, &sq->fence, NULL);
791       break;
792    case PIPE_QUERY_PRIMITIVES_GENERATED:
793    case PIPE_QUERY_PRIMITIVES_EMITTED:
794    case PIPE_QUERY_SO_STATISTICS:
795    case PIPE_QUERY_TIMESTAMP:
796       assert(svga_have_vgpu10(svga));
797       destroy_query_vgpu10(svga, sq);
798       sws->fence_reference(sws, &sq->fence, NULL);
799       break;
800    case SVGA_QUERY_NUM_DRAW_CALLS:
801    case SVGA_QUERY_NUM_FALLBACKS:
802    case SVGA_QUERY_NUM_FLUSHES:
803    case SVGA_QUERY_NUM_VALIDATIONS:
804    case SVGA_QUERY_MAP_BUFFER_TIME:
805    case SVGA_QUERY_NUM_BUFFERS_MAPPED:
806    case SVGA_QUERY_NUM_TEXTURES_MAPPED:
807    case SVGA_QUERY_NUM_BYTES_UPLOADED:
808    case SVGA_QUERY_NUM_COMMAND_BUFFERS:
809    case SVGA_QUERY_COMMAND_BUFFER_SIZE:
810    case SVGA_QUERY_FLUSH_TIME:
811    case SVGA_QUERY_SURFACE_WRITE_FLUSHES:
812    case SVGA_QUERY_MEMORY_USED:
813    case SVGA_QUERY_NUM_SHADERS:
814    case SVGA_QUERY_NUM_RESOURCES:
815    case SVGA_QUERY_NUM_STATE_OBJECTS:
816    case SVGA_QUERY_NUM_SURFACE_VIEWS:
817    case SVGA_QUERY_NUM_GENERATE_MIPMAP:
818    case SVGA_QUERY_NUM_READBACKS:
819    case SVGA_QUERY_NUM_RESOURCE_UPDATES:
820    case SVGA_QUERY_NUM_BUFFER_UPLOADS:
821    case SVGA_QUERY_NUM_CONST_BUF_UPDATES:
822    case SVGA_QUERY_NUM_CONST_UPDATES:
823    case SVGA_QUERY_NUM_FAILED_ALLOCATIONS:
824    case SVGA_QUERY_NUM_COMMANDS_PER_DRAW:
825    case SVGA_QUERY_NUM_SHADER_RELOCATIONS:
826    case SVGA_QUERY_NUM_SURFACE_RELOCATIONS:
827    case SVGA_QUERY_SHADER_MEM_USED:
828       /* nothing */
829       break;
830    default:
831       assert(!"svga: unexpected query type in svga_destroy_query()");
832    }
833 
834    /* Free the query id */
835    util_bitmask_clear(svga->query_id_bm, sq->id);
836 
837    FREE(sq);
838 }
839 
840 
841 static bool
svga_begin_query(struct pipe_context * pipe,struct pipe_query * q)842 svga_begin_query(struct pipe_context *pipe, struct pipe_query *q)
843 {
844    struct svga_context *svga = svga_context(pipe);
845    struct svga_query *sq = svga_query(q);
846    enum pipe_error ret = PIPE_OK;
847 
848    assert(sq);
849    assert(sq->type < SVGA_QUERY_MAX);
850 
851    /* Need to flush out buffered drawing commands so that they don't
852     * get counted in the query results.
853     */
854    svga_hwtnl_flush_retry(svga);
855 
856    switch (sq->type) {
857    case PIPE_QUERY_OCCLUSION_COUNTER:
858    case PIPE_QUERY_OCCLUSION_PREDICATE:
859    case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
860       if (svga_have_vgpu10(svga)) {
861          ret = begin_query_vgpu10(svga, sq);
862          /* also need to start the associated occlusion predicate query */
863          if (sq->predicate) {
864             enum pipe_error status;
865             status = begin_query_vgpu10(svga, svga_query(sq->predicate));
866             assert(status == PIPE_OK);
867             (void) status;
868          }
869       } else {
870          begin_query_vgpu9(svga, sq);
871       }
872       assert(ret == PIPE_OK);
873       (void) ret;
874       break;
875    case PIPE_QUERY_PRIMITIVES_GENERATED:
876    case PIPE_QUERY_PRIMITIVES_EMITTED:
877    case PIPE_QUERY_SO_STATISTICS:
878    case PIPE_QUERY_TIMESTAMP:
879       assert(svga_have_vgpu10(svga));
880       ret = begin_query_vgpu10(svga, sq);
881       assert(ret == PIPE_OK);
882       break;
883    case SVGA_QUERY_NUM_DRAW_CALLS:
884       sq->begin_count = svga->hud.num_draw_calls;
885       break;
886    case SVGA_QUERY_NUM_FALLBACKS:
887       sq->begin_count = svga->hud.num_fallbacks;
888       break;
889    case SVGA_QUERY_NUM_FLUSHES:
890       sq->begin_count = svga->hud.num_flushes;
891       break;
892    case SVGA_QUERY_NUM_VALIDATIONS:
893       sq->begin_count = svga->hud.num_validations;
894       break;
895    case SVGA_QUERY_MAP_BUFFER_TIME:
896       sq->begin_count = svga->hud.map_buffer_time;
897       break;
898    case SVGA_QUERY_NUM_BUFFERS_MAPPED:
899       sq->begin_count = svga->hud.num_buffers_mapped;
900       break;
901    case SVGA_QUERY_NUM_TEXTURES_MAPPED:
902       sq->begin_count = svga->hud.num_textures_mapped;
903       break;
904    case SVGA_QUERY_NUM_BYTES_UPLOADED:
905       sq->begin_count = svga->hud.num_bytes_uploaded;
906       break;
907    case SVGA_QUERY_NUM_COMMAND_BUFFERS:
908       sq->begin_count = svga->swc->num_command_buffers;
909       break;
910    case SVGA_QUERY_COMMAND_BUFFER_SIZE:
911       sq->begin_count = svga->hud.command_buffer_size;
912       break;
913    case SVGA_QUERY_FLUSH_TIME:
914       sq->begin_count = svga->hud.flush_time;
915       break;
916    case SVGA_QUERY_SURFACE_WRITE_FLUSHES:
917       sq->begin_count = svga->hud.surface_write_flushes;
918       break;
919    case SVGA_QUERY_NUM_READBACKS:
920       sq->begin_count = svga->hud.num_readbacks;
921       break;
922    case SVGA_QUERY_NUM_RESOURCE_UPDATES:
923       sq->begin_count = svga->hud.num_resource_updates;
924       break;
925    case SVGA_QUERY_NUM_BUFFER_UPLOADS:
926       sq->begin_count = svga->hud.num_buffer_uploads;
927       break;
928    case SVGA_QUERY_NUM_CONST_BUF_UPDATES:
929       sq->begin_count = svga->hud.num_const_buf_updates;
930       break;
931    case SVGA_QUERY_NUM_CONST_UPDATES:
932       sq->begin_count = svga->hud.num_const_updates;
933       break;
934    case SVGA_QUERY_NUM_SHADER_RELOCATIONS:
935       sq->begin_count = svga->swc->num_shader_reloc;
936       break;
937    case SVGA_QUERY_NUM_SURFACE_RELOCATIONS:
938       sq->begin_count = svga->swc->num_surf_reloc;
939       break;
940    case SVGA_QUERY_MEMORY_USED:
941    case SVGA_QUERY_NUM_SHADERS:
942    case SVGA_QUERY_NUM_RESOURCES:
943    case SVGA_QUERY_NUM_STATE_OBJECTS:
944    case SVGA_QUERY_NUM_SURFACE_VIEWS:
945    case SVGA_QUERY_NUM_GENERATE_MIPMAP:
946    case SVGA_QUERY_NUM_FAILED_ALLOCATIONS:
947    case SVGA_QUERY_NUM_COMMANDS_PER_DRAW:
948    case SVGA_QUERY_SHADER_MEM_USED:
949       /* nothing */
950       break;
951    default:
952       assert(!"unexpected query type in svga_begin_query()");
953    }
954 
955    SVGA_DBG(DEBUG_QUERY, "%s sq=0x%x id=%d type=%d svga_type=%d\n",
956             __FUNCTION__, sq, sq->id, sq->type, sq->svga_type);
957 
958    sq->active = TRUE;
959 
960    return true;
961 }
962 
963 
964 static bool
svga_end_query(struct pipe_context * pipe,struct pipe_query * q)965 svga_end_query(struct pipe_context *pipe, struct pipe_query *q)
966 {
967    struct svga_context *svga = svga_context(pipe);
968    struct svga_query *sq = svga_query(q);
969 
970    assert(sq);
971    assert(sq->type < SVGA_QUERY_MAX);
972 
973    SVGA_DBG(DEBUG_QUERY, "%s sq=0x%x type=%d\n",
974             __FUNCTION__, sq, sq->type);
975 
976    if (sq->type == PIPE_QUERY_TIMESTAMP && !sq->active)
977       svga_begin_query(pipe, q);
978 
979    SVGA_DBG(DEBUG_QUERY, "%s sq=0x%x id=%d type=%d svga_type=%d\n",
980             __FUNCTION__, sq, sq->id, sq->type, sq->svga_type);
981 
982    svga_hwtnl_flush_retry(svga);
983 
984    assert(sq->active);
985 
986    switch (sq->type) {
987    case PIPE_QUERY_OCCLUSION_COUNTER:
988    case PIPE_QUERY_OCCLUSION_PREDICATE:
989    case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE:
990       if (svga_have_vgpu10(svga)) {
991          end_query_vgpu10(svga, sq);
992          /* also need to end the associated occlusion predicate query */
993          if (sq->predicate) {
994             end_query_vgpu10(svga, svga_query(sq->predicate));
995          }
996       } else {
997          end_query_vgpu9(svga, sq);
998       }
999       break;
1000    case PIPE_QUERY_PRIMITIVES_GENERATED:
1001    case PIPE_QUERY_PRIMITIVES_EMITTED:
1002    case PIPE_QUERY_SO_STATISTICS:
1003    case PIPE_QUERY_TIMESTAMP:
1004       assert(svga_have_vgpu10(svga));
1005       end_query_vgpu10(svga, sq);
1006       break;
1007    case SVGA_QUERY_NUM_DRAW_CALLS:
1008       sq->end_count = svga->hud.num_draw_calls;
1009       break;
1010    case SVGA_QUERY_NUM_FALLBACKS:
1011       sq->end_count = svga->hud.num_fallbacks;
1012       break;
1013    case SVGA_QUERY_NUM_FLUSHES:
1014       sq->end_count = svga->hud.num_flushes;
1015       break;
1016    case SVGA_QUERY_NUM_VALIDATIONS:
1017       sq->end_count = svga->hud.num_validations;
1018       break;
1019    case SVGA_QUERY_MAP_BUFFER_TIME:
1020       sq->end_count = svga->hud.map_buffer_time;
1021       break;
1022    case SVGA_QUERY_NUM_BUFFERS_MAPPED:
1023       sq->end_count = svga->hud.num_buffers_mapped;
1024       break;
1025    case SVGA_QUERY_NUM_TEXTURES_MAPPED:
1026       sq->end_count = svga->hud.num_textures_mapped;
1027       break;
1028    case SVGA_QUERY_NUM_BYTES_UPLOADED:
1029       sq->end_count = svga->hud.num_bytes_uploaded;
1030       break;
1031    case SVGA_QUERY_NUM_COMMAND_BUFFERS:
1032       sq->end_count = svga->swc->num_command_buffers;
1033       break;
1034    case SVGA_QUERY_COMMAND_BUFFER_SIZE:
1035       sq->end_count = svga->hud.command_buffer_size;
1036       break;
1037    case SVGA_QUERY_FLUSH_TIME:
1038       sq->end_count = svga->hud.flush_time;
1039       break;
1040    case SVGA_QUERY_SURFACE_WRITE_FLUSHES:
1041       sq->end_count = svga->hud.surface_write_flushes;
1042       break;
1043    case SVGA_QUERY_NUM_READBACKS:
1044       sq->end_count = svga->hud.num_readbacks;
1045       break;
1046    case SVGA_QUERY_NUM_RESOURCE_UPDATES:
1047       sq->end_count = svga->hud.num_resource_updates;
1048       break;
1049    case SVGA_QUERY_NUM_BUFFER_UPLOADS:
1050       sq->end_count = svga->hud.num_buffer_uploads;
1051       break;
1052    case SVGA_QUERY_NUM_CONST_BUF_UPDATES:
1053       sq->end_count = svga->hud.num_const_buf_updates;
1054       break;
1055    case SVGA_QUERY_NUM_CONST_UPDATES:
1056       sq->end_count = svga->hud.num_const_updates;
1057       break;
1058    case SVGA_QUERY_NUM_SHADER_RELOCATIONS:
1059       sq->end_count = svga->swc->num_shader_reloc;
1060       break;
1061    case SVGA_QUERY_NUM_SURFACE_RELOCATIONS:
1062       sq->end_count = svga->swc->num_surf_reloc;
1063       break;
1064    case SVGA_QUERY_MEMORY_USED:
1065    case SVGA_QUERY_NUM_SHADERS:
1066    case SVGA_QUERY_NUM_RESOURCES:
1067    case SVGA_QUERY_NUM_STATE_OBJECTS:
1068    case SVGA_QUERY_NUM_SURFACE_VIEWS:
1069    case SVGA_QUERY_NUM_GENERATE_MIPMAP:
1070    case SVGA_QUERY_NUM_FAILED_ALLOCATIONS:
1071    case SVGA_QUERY_NUM_COMMANDS_PER_DRAW:
1072    case SVGA_QUERY_SHADER_MEM_USED:
1073       /* nothing */
1074       break;
1075    default:
1076       assert(!"unexpected query type in svga_end_query()");
1077    }
1078    sq->active = FALSE;
1079    return true;
1080 }
1081 
1082 
1083 static bool
svga_get_query_result(struct pipe_context * pipe,struct pipe_query * q,bool wait,union pipe_query_result * vresult)1084 svga_get_query_result(struct pipe_context *pipe,
1085                       struct pipe_query *q,
1086                       bool wait,
1087                       union pipe_query_result *vresult)
1088 {
1089    struct svga_screen *svgascreen = svga_screen(pipe->screen);
1090    struct svga_context *svga = svga_context(pipe);
1091    struct svga_query *sq = svga_query(q);
1092    uint64_t *result = (uint64_t *)vresult;
1093    bool ret = true;
1094 
1095    assert(sq);
1096 
1097    SVGA_DBG(DEBUG_QUERY, "%s sq=0x%x id=%d wait: %d\n",
1098             __FUNCTION__, sq, sq->id, wait);
1099 
1100    switch (sq->type) {
1101    case PIPE_QUERY_OCCLUSION_COUNTER:
1102       if (svga_have_vgpu10(svga)) {
1103          SVGADXOcclusionQueryResult occResult;
1104          ret = get_query_result_vgpu10(svga, sq, wait,
1105                                        (void *)&occResult, sizeof(occResult));
1106          *result = (uint64_t)occResult.samplesRendered;
1107       } else {
1108          ret = get_query_result_vgpu9(svga, sq, wait, result);
1109       }
1110       break;
1111    case PIPE_QUERY_OCCLUSION_PREDICATE:
1112    case PIPE_QUERY_OCCLUSION_PREDICATE_CONSERVATIVE: {
1113       if (svga_have_vgpu10(svga)) {
1114          SVGADXOcclusionPredicateQueryResult occResult;
1115          ret = get_query_result_vgpu10(svga, sq, wait,
1116                                        (void *)&occResult, sizeof(occResult));
1117          vresult->b = occResult.anySamplesRendered != 0;
1118       } else {
1119          uint64_t count = 0;
1120          ret = get_query_result_vgpu9(svga, sq, wait, &count);
1121          vresult->b = count != 0;
1122       }
1123       break;
1124    }
1125    case PIPE_QUERY_SO_STATISTICS: {
1126       SVGADXStreamOutStatisticsQueryResult sResult;
1127       struct pipe_query_data_so_statistics *pResult =
1128          (struct pipe_query_data_so_statistics *)vresult;
1129 
1130       assert(svga_have_vgpu10(svga));
1131       ret = get_query_result_vgpu10(svga, sq, wait,
1132                                     (void *)&sResult, sizeof(sResult));
1133       pResult->num_primitives_written = sResult.numPrimitivesWritten;
1134       pResult->primitives_storage_needed = sResult.numPrimitivesRequired;
1135       break;
1136    }
1137    case PIPE_QUERY_TIMESTAMP: {
1138       SVGADXTimestampQueryResult sResult;
1139 
1140       assert(svga_have_vgpu10(svga));
1141       ret = get_query_result_vgpu10(svga, sq, wait,
1142                                     (void *)&sResult, sizeof(sResult));
1143       *result = (uint64_t)sResult.timestamp;
1144       break;
1145    }
1146    case PIPE_QUERY_PRIMITIVES_GENERATED: {
1147       SVGADXStreamOutStatisticsQueryResult sResult;
1148 
1149       assert(svga_have_vgpu10(svga));
1150       ret = get_query_result_vgpu10(svga, sq, wait,
1151                                     (void *)&sResult, sizeof sResult);
1152       *result = (uint64_t)sResult.numPrimitivesRequired;
1153       break;
1154    }
1155    case PIPE_QUERY_PRIMITIVES_EMITTED: {
1156       SVGADXStreamOutStatisticsQueryResult sResult;
1157 
1158       assert(svga_have_vgpu10(svga));
1159       ret = get_query_result_vgpu10(svga, sq, wait,
1160                                     (void *)&sResult, sizeof sResult);
1161       *result = (uint64_t)sResult.numPrimitivesWritten;
1162       break;
1163    }
1164    /* These are per-frame counters */
1165    case SVGA_QUERY_NUM_DRAW_CALLS:
1166    case SVGA_QUERY_NUM_FALLBACKS:
1167    case SVGA_QUERY_NUM_FLUSHES:
1168    case SVGA_QUERY_NUM_VALIDATIONS:
1169    case SVGA_QUERY_MAP_BUFFER_TIME:
1170    case SVGA_QUERY_NUM_BUFFERS_MAPPED:
1171    case SVGA_QUERY_NUM_TEXTURES_MAPPED:
1172    case SVGA_QUERY_NUM_BYTES_UPLOADED:
1173    case SVGA_QUERY_NUM_COMMAND_BUFFERS:
1174    case SVGA_QUERY_COMMAND_BUFFER_SIZE:
1175    case SVGA_QUERY_FLUSH_TIME:
1176    case SVGA_QUERY_SURFACE_WRITE_FLUSHES:
1177    case SVGA_QUERY_NUM_READBACKS:
1178    case SVGA_QUERY_NUM_RESOURCE_UPDATES:
1179    case SVGA_QUERY_NUM_BUFFER_UPLOADS:
1180    case SVGA_QUERY_NUM_CONST_BUF_UPDATES:
1181    case SVGA_QUERY_NUM_CONST_UPDATES:
1182    case SVGA_QUERY_NUM_SHADER_RELOCATIONS:
1183    case SVGA_QUERY_NUM_SURFACE_RELOCATIONS:
1184       vresult->u64 = sq->end_count - sq->begin_count;
1185       break;
1186    /* These are running total counters */
1187    case SVGA_QUERY_MEMORY_USED:
1188       vresult->u64 = svgascreen->hud.total_resource_bytes;
1189       break;
1190    case SVGA_QUERY_NUM_SHADERS:
1191       vresult->u64 = svga->hud.num_shaders;
1192       break;
1193    case SVGA_QUERY_NUM_RESOURCES:
1194       vresult->u64 = svgascreen->hud.num_resources;
1195       break;
1196    case SVGA_QUERY_NUM_STATE_OBJECTS:
1197       vresult->u64 = (svga->hud.num_blend_objects +
1198                       svga->hud.num_depthstencil_objects +
1199                       svga->hud.num_rasterizer_objects +
1200                       svga->hud.num_sampler_objects +
1201                       svga->hud.num_samplerview_objects +
1202                       svga->hud.num_vertexelement_objects);
1203       break;
1204    case SVGA_QUERY_NUM_SURFACE_VIEWS:
1205       vresult->u64 = svga->hud.num_surface_views;
1206       break;
1207    case SVGA_QUERY_NUM_GENERATE_MIPMAP:
1208       vresult->u64 = svga->hud.num_generate_mipmap;
1209       break;
1210    case SVGA_QUERY_NUM_FAILED_ALLOCATIONS:
1211       vresult->u64 = svgascreen->hud.num_failed_allocations;
1212       break;
1213    case SVGA_QUERY_NUM_COMMANDS_PER_DRAW:
1214       vresult->f = (float) svga->swc->num_commands
1215          / (float) svga->swc->num_draw_commands;
1216       break;
1217    case SVGA_QUERY_SHADER_MEM_USED:
1218       vresult->u64 = svga->hud.shader_mem_used;
1219       break;
1220    default:
1221       assert(!"unexpected query type in svga_get_query_result");
1222    }
1223 
1224    SVGA_DBG(DEBUG_QUERY, "%s result %d\n", __FUNCTION__, *((uint64_t *)vresult));
1225 
1226    return ret;
1227 }
1228 
1229 static void
svga_render_condition(struct pipe_context * pipe,struct pipe_query * q,bool condition,enum pipe_render_cond_flag mode)1230 svga_render_condition(struct pipe_context *pipe, struct pipe_query *q,
1231                       bool condition, enum pipe_render_cond_flag mode)
1232 {
1233    struct svga_context *svga = svga_context(pipe);
1234    struct svga_winsys_screen *sws = svga_screen(svga->pipe.screen)->sws;
1235    struct svga_query *sq = svga_query(q);
1236    SVGA3dQueryId queryId;
1237 
1238    SVGA_DBG(DEBUG_QUERY, "%s\n", __FUNCTION__);
1239 
1240    assert(svga_have_vgpu10(svga));
1241    if (sq == NULL) {
1242       queryId = SVGA3D_INVALID_ID;
1243    }
1244    else {
1245       assert(sq->svga_type == SVGA3D_QUERYTYPE_OCCLUSION ||
1246              sq->svga_type == SVGA3D_QUERYTYPE_OCCLUSIONPREDICATE);
1247 
1248       if (sq->svga_type == SVGA3D_QUERYTYPE_OCCLUSION) {
1249          assert(sq->predicate);
1250          /**
1251           * For conditional rendering, make sure to use the associated
1252           * predicate query.
1253           */
1254          sq = svga_query(sq->predicate);
1255       }
1256       queryId = sq->id;
1257 
1258       if ((mode == PIPE_RENDER_COND_WAIT ||
1259            mode == PIPE_RENDER_COND_BY_REGION_WAIT) && sq->fence) {
1260          sws->fence_finish(sws, sq->fence, PIPE_TIMEOUT_INFINITE,
1261                            SVGA_FENCE_FLAG_QUERY);
1262       }
1263    }
1264    /*
1265     * if the kernel module doesn't support the predication command,
1266     * we'll just render unconditionally.
1267     * This is probably acceptable for the typical case of occlusion culling.
1268     */
1269    if (sws->have_set_predication_cmd) {
1270       SVGA_RETRY(svga, SVGA3D_vgpu10_SetPredication(svga->swc, queryId,
1271                                                     (uint32) condition));
1272       svga->pred.query_id = queryId;
1273       svga->pred.cond = condition;
1274    }
1275 
1276    svga->render_condition = (sq != NULL);
1277 }
1278 
1279 
1280 /*
1281  * This function is a workaround because we lack the ability to query
1282  * renderer's time synchornously.
1283  */
1284 static uint64_t
svga_get_timestamp(struct pipe_context * pipe)1285 svga_get_timestamp(struct pipe_context *pipe)
1286 {
1287    struct pipe_query *q = svga_create_query(pipe, PIPE_QUERY_TIMESTAMP, 0);
1288    union pipe_query_result result;
1289 
1290    svga_begin_query(pipe, q);
1291    svga_end_query(pipe,q);
1292    svga_get_query_result(pipe, q, TRUE, &result);
1293    svga_destroy_query(pipe, q);
1294 
1295    return result.u64;
1296 }
1297 
1298 
1299 static void
svga_set_active_query_state(struct pipe_context * pipe,bool enable)1300 svga_set_active_query_state(struct pipe_context *pipe, bool enable)
1301 {
1302 }
1303 
1304 
1305 /**
1306  * \brief Toggle conditional rendering if already enabled
1307  *
1308  * \param svga[in]  The svga context
1309  * \param render_condition_enabled[in]  Whether to ignore requests to turn
1310  * conditional rendering off
1311  * \param on[in]  Whether to turn conditional rendering on or off
1312  */
1313 void
svga_toggle_render_condition(struct svga_context * svga,boolean render_condition_enabled,boolean on)1314 svga_toggle_render_condition(struct svga_context *svga,
1315                              boolean render_condition_enabled,
1316                              boolean on)
1317 {
1318    SVGA3dQueryId query_id;
1319 
1320    if (render_condition_enabled ||
1321        svga->pred.query_id == SVGA3D_INVALID_ID) {
1322       return;
1323    }
1324 
1325    /*
1326     * If we get here, it means that the system supports
1327     * conditional rendering since svga->pred.query_id has already been
1328     * modified for this context and thus support has already been
1329     * verified.
1330     */
1331    query_id = on ? svga->pred.query_id : SVGA3D_INVALID_ID;
1332 
1333    SVGA_RETRY(svga, SVGA3D_vgpu10_SetPredication(svga->swc, query_id,
1334                                                  (uint32) svga->pred.cond));
1335 }
1336 
1337 
1338 void
svga_init_query_functions(struct svga_context * svga)1339 svga_init_query_functions(struct svga_context *svga)
1340 {
1341    svga->pipe.create_query = svga_create_query;
1342    svga->pipe.destroy_query = svga_destroy_query;
1343    svga->pipe.begin_query = svga_begin_query;
1344    svga->pipe.end_query = svga_end_query;
1345    svga->pipe.get_query_result = svga_get_query_result;
1346    svga->pipe.set_active_query_state = svga_set_active_query_state;
1347    svga->pipe.render_condition = svga_render_condition;
1348    svga->pipe.get_timestamp = svga_get_timestamp;
1349 }
1350