• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /****************************************************************************
2  * Copyright (C) 2015 Intel Corporation.   All Rights Reserved.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8  * and/or sell copies of the Software, and to permit persons to whom the
9  * Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice (including the next
12  * paragraph) shall be included in all copies or substantial portions of the
13  * Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
18  * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19  * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20  * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21  * IN THE SOFTWARE.
22  ***************************************************************************/
23 
24 #include "swr_context.h"
25 #include "swr_memory.h"
26 #include "swr_screen.h"
27 #include "swr_resource.h"
28 #include "swr_scratch.h"
29 #include "swr_query.h"
30 #include "swr_fence.h"
31 
32 #include "util/u_memory.h"
33 #include "util/u_inlines.h"
34 #include "util/format/u_format.h"
35 #include "util/u_atomic.h"
36 #include "util/u_upload_mgr.h"
37 #include "util/u_transfer.h"
38 #include "util/u_surface.h"
39 
40 #include "api.h"
41 #include "backend.h"
42 #include "knobs.h"
43 
44 static struct pipe_surface *
swr_create_surface(struct pipe_context * pipe,struct pipe_resource * pt,const struct pipe_surface * surf_tmpl)45 swr_create_surface(struct pipe_context *pipe,
46                    struct pipe_resource *pt,
47                    const struct pipe_surface *surf_tmpl)
48 {
49    struct pipe_surface *ps;
50 
51    ps = CALLOC_STRUCT(pipe_surface);
52    if (ps) {
53       pipe_reference_init(&ps->reference, 1);
54       pipe_resource_reference(&ps->texture, pt);
55       ps->context = pipe;
56       ps->format = surf_tmpl->format;
57       if (pt->target != PIPE_BUFFER) {
58          assert(surf_tmpl->u.tex.level <= pt->last_level);
59          ps->width = u_minify(pt->width0, surf_tmpl->u.tex.level);
60          ps->height = u_minify(pt->height0, surf_tmpl->u.tex.level);
61          ps->u.tex.level = surf_tmpl->u.tex.level;
62          ps->u.tex.first_layer = surf_tmpl->u.tex.first_layer;
63          ps->u.tex.last_layer = surf_tmpl->u.tex.last_layer;
64       } else {
65          /* setting width as number of elements should get us correct
66           * renderbuffer width */
67          ps->width = surf_tmpl->u.buf.last_element
68             - surf_tmpl->u.buf.first_element + 1;
69          ps->height = pt->height0;
70          ps->u.buf.first_element = surf_tmpl->u.buf.first_element;
71          ps->u.buf.last_element = surf_tmpl->u.buf.last_element;
72          assert(ps->u.buf.first_element <= ps->u.buf.last_element);
73          assert(ps->u.buf.last_element < ps->width);
74       }
75    }
76    return ps;
77 }
78 
79 static void
swr_surface_destroy(struct pipe_context * pipe,struct pipe_surface * surf)80 swr_surface_destroy(struct pipe_context *pipe, struct pipe_surface *surf)
81 {
82    assert(surf->texture);
83    struct pipe_resource *resource = surf->texture;
84 
85    /* If the resource has been drawn to, store tiles. */
86    swr_store_dirty_resource(pipe, resource, SWR_TILE_RESOLVED);
87 
88    pipe_resource_reference(&resource, NULL);
89    FREE(surf);
90 }
91 
92 
93 static void *
swr_transfer_map(struct pipe_context * pipe,struct pipe_resource * resource,unsigned level,unsigned usage,const struct pipe_box * box,struct pipe_transfer ** transfer)94 swr_transfer_map(struct pipe_context *pipe,
95                  struct pipe_resource *resource,
96                  unsigned level,
97                  unsigned usage,
98                  const struct pipe_box *box,
99                  struct pipe_transfer **transfer)
100 {
101    struct swr_screen *screen = swr_screen(pipe->screen);
102    struct swr_resource *spr = swr_resource(resource);
103    struct pipe_transfer *pt;
104    enum pipe_format format = resource->format;
105 
106    assert(resource);
107    assert(level <= resource->last_level);
108 
109    /* If mapping an attached rendertarget, store tiles to surface and set
110     * postStoreTileState to SWR_TILE_INVALID so tiles get reloaded on next use
111     * and nothing needs to be done at unmap. */
112    swr_store_dirty_resource(pipe, resource, SWR_TILE_INVALID);
113 
114    if (!(usage & PIPE_MAP_UNSYNCHRONIZED)) {
115       /* If resource is in use, finish fence before mapping.
116        * Unless requested not to block, then if not done return NULL map */
117       if (usage & PIPE_MAP_DONTBLOCK) {
118          if (swr_is_fence_pending(screen->flush_fence))
119             return NULL;
120       } else {
121          if (spr->status) {
122             /* But, if there's no fence pending, submit one.
123              * XXX: Remove once draw timestamps are finished. */
124             if (!swr_is_fence_pending(screen->flush_fence))
125                swr_fence_submit(swr_context(pipe), screen->flush_fence);
126 
127             swr_fence_finish(pipe->screen, NULL, screen->flush_fence, 0);
128             swr_resource_unused(resource);
129          }
130       }
131    }
132 
133    pt = CALLOC_STRUCT(pipe_transfer);
134    if (!pt)
135       return NULL;
136    pipe_resource_reference(&pt->resource, resource);
137    pt->usage = (pipe_map_flags)usage;
138    pt->level = level;
139    pt->box = *box;
140    pt->stride = spr->swr.pitch;
141    pt->layer_stride = spr->swr.qpitch * spr->swr.pitch;
142 
143    /* if we're mapping the depth/stencil, copy in stencil for the section
144     * being read in
145     */
146    if (usage & PIPE_MAP_READ && spr->has_depth && spr->has_stencil) {
147       size_t zbase, sbase;
148       for (int z = box->z; z < box->z + box->depth; z++) {
149          zbase = (z * spr->swr.qpitch + box->y) * spr->swr.pitch +
150             spr->mip_offsets[level];
151          sbase = (z * spr->secondary.qpitch + box->y) * spr->secondary.pitch +
152             spr->secondary_mip_offsets[level];
153          for (int y = box->y; y < box->y + box->height; y++) {
154             if (spr->base.format == PIPE_FORMAT_Z24_UNORM_S8_UINT) {
155                for (int x = box->x; x < box->x + box->width; x++)
156                   ((uint8_t*)(spr->swr.xpBaseAddress))[zbase + 4 * x + 3] =
157                      ((uint8_t*)(spr->secondary.xpBaseAddress))[sbase + x];
158             } else if (spr->base.format == PIPE_FORMAT_Z32_FLOAT_S8X24_UINT) {
159                for (int x = box->x; x < box->x + box->width; x++)
160                   ((uint8_t*)(spr->swr.xpBaseAddress))[zbase + 8 * x + 4] =
161                      ((uint8_t*)(spr->secondary.xpBaseAddress))[sbase + x];
162             }
163             zbase += spr->swr.pitch;
164             sbase += spr->secondary.pitch;
165          }
166       }
167    }
168 
169    unsigned offset = box->z * pt->layer_stride +
170       util_format_get_nblocksy(format, box->y) * pt->stride +
171       util_format_get_stride(format, box->x);
172 
173    *transfer = pt;
174 
175    return (void*)(spr->swr.xpBaseAddress + offset + spr->mip_offsets[level]);
176 }
177 
178 static void
swr_transfer_flush_region(struct pipe_context * pipe,struct pipe_transfer * transfer,const struct pipe_box * flush_box)179 swr_transfer_flush_region(struct pipe_context *pipe,
180                           struct pipe_transfer *transfer,
181                           const struct pipe_box *flush_box)
182 {
183    assert(transfer->resource);
184    assert(transfer->usage & PIPE_MAP_WRITE);
185 
186    struct swr_resource *spr = swr_resource(transfer->resource);
187    if (!spr->has_depth || !spr->has_stencil)
188       return;
189 
190    size_t zbase, sbase;
191    struct pipe_box box = *flush_box;
192    box.x += transfer->box.x;
193    box.y += transfer->box.y;
194    box.z += transfer->box.z;
195    for (int z = box.z; z < box.z + box.depth; z++) {
196       zbase = (z * spr->swr.qpitch + box.y) * spr->swr.pitch +
197          spr->mip_offsets[transfer->level];
198       sbase = (z * spr->secondary.qpitch + box.y) * spr->secondary.pitch +
199          spr->secondary_mip_offsets[transfer->level];
200       for (int y = box.y; y < box.y + box.height; y++) {
201          if (spr->base.format == PIPE_FORMAT_Z24_UNORM_S8_UINT) {
202             for (int x = box.x; x < box.x + box.width; x++)
203                ((uint8_t*)(spr->secondary.xpBaseAddress))[sbase + x] =
204                   ((uint8_t*)(spr->swr.xpBaseAddress))[zbase + 4 * x + 3];
205          } else if (spr->base.format == PIPE_FORMAT_Z32_FLOAT_S8X24_UINT) {
206             for (int x = box.x; x < box.x + box.width; x++)
207                ((uint8_t*)(spr->secondary.xpBaseAddress))[sbase + x] =
208                   ((uint8_t*)(spr->swr.xpBaseAddress))[zbase + 8 * x + 4];
209          }
210          zbase += spr->swr.pitch;
211          sbase += spr->secondary.pitch;
212       }
213    }
214 }
215 
216 static void
swr_transfer_unmap(struct pipe_context * pipe,struct pipe_transfer * transfer)217 swr_transfer_unmap(struct pipe_context *pipe, struct pipe_transfer *transfer)
218 {
219    assert(transfer->resource);
220 
221    struct swr_resource *spr = swr_resource(transfer->resource);
222    /* if we're mapping the depth/stencil, copy in stencil for the section
223     * being written out
224     */
225    if (transfer->usage & PIPE_MAP_WRITE &&
226        !(transfer->usage & PIPE_MAP_FLUSH_EXPLICIT) &&
227        spr->has_depth && spr->has_stencil) {
228       struct pipe_box box;
229       u_box_3d(0, 0, 0, transfer->box.width, transfer->box.height,
230                transfer->box.depth, &box);
231       swr_transfer_flush_region(pipe, transfer, &box);
232    }
233 
234    pipe_resource_reference(&transfer->resource, NULL);
235    FREE(transfer);
236 }
237 
238 
239 static void
swr_resource_copy(struct pipe_context * pipe,struct pipe_resource * dst,unsigned dst_level,unsigned dstx,unsigned dsty,unsigned dstz,struct pipe_resource * src,unsigned src_level,const struct pipe_box * src_box)240 swr_resource_copy(struct pipe_context *pipe,
241                   struct pipe_resource *dst,
242                   unsigned dst_level,
243                   unsigned dstx,
244                   unsigned dsty,
245                   unsigned dstz,
246                   struct pipe_resource *src,
247                   unsigned src_level,
248                   const struct pipe_box *src_box)
249 {
250    struct swr_screen *screen = swr_screen(pipe->screen);
251 
252    /* If either the src or dst is a renderTarget, store tiles before copy */
253    swr_store_dirty_resource(pipe, src, SWR_TILE_RESOLVED);
254    swr_store_dirty_resource(pipe, dst, SWR_TILE_RESOLVED);
255 
256    swr_fence_finish(pipe->screen, NULL, screen->flush_fence, 0);
257    swr_resource_unused(src);
258    swr_resource_unused(dst);
259 
260    if ((dst->target == PIPE_BUFFER && src->target == PIPE_BUFFER)
261        || (dst->target != PIPE_BUFFER && src->target != PIPE_BUFFER)) {
262       util_resource_copy_region(
263          pipe, dst, dst_level, dstx, dsty, dstz, src, src_level, src_box);
264       return;
265    }
266 
267    debug_printf("unhandled swr_resource_copy\n");
268 }
269 
270 
271 static void
swr_blit(struct pipe_context * pipe,const struct pipe_blit_info * blit_info)272 swr_blit(struct pipe_context *pipe, const struct pipe_blit_info *blit_info)
273 {
274    struct swr_context *ctx = swr_context(pipe);
275    /* Make a copy of the const blit_info, so we can modify it */
276    struct pipe_blit_info info = *blit_info;
277 
278    if (info.render_condition_enable && !swr_check_render_cond(pipe))
279       return;
280 
281    if (info.src.resource->nr_samples > 1 && info.dst.resource->nr_samples <= 1
282        && !util_format_is_depth_or_stencil(info.src.resource->format)
283        && !util_format_is_pure_integer(info.src.resource->format)) {
284       debug_printf("swr_blit: color resolve : %d -> %d\n",
285             info.src.resource->nr_samples, info.dst.resource->nr_samples);
286 
287       /* Resolve is done as part of the surface store. */
288       swr_store_dirty_resource(pipe, info.src.resource, SWR_TILE_RESOLVED);
289 
290       struct pipe_resource *src_resource = info.src.resource;
291       struct pipe_resource *resolve_target =
292          swr_resource(src_resource)->resolve_target;
293 
294       /* The resolve target becomes the new source for the blit. */
295       info.src.resource = resolve_target;
296    }
297 
298    if (util_try_blit_via_copy_region(pipe, &info)) {
299       return; /* done */
300    }
301 
302    if (info.mask & PIPE_MASK_S) {
303       debug_printf("swr: cannot blit stencil, skipping\n");
304       info.mask &= ~PIPE_MASK_S;
305    }
306 
307    if (!util_blitter_is_blit_supported(ctx->blitter, &info)) {
308       debug_printf("swr: blit unsupported %s -> %s\n",
309                    util_format_short_name(info.src.resource->format),
310                    util_format_short_name(info.dst.resource->format));
311       return;
312    }
313 
314    if (ctx->active_queries) {
315       ctx->api.pfnSwrEnableStatsFE(ctx->swrContext, FALSE);
316       ctx->api.pfnSwrEnableStatsBE(ctx->swrContext, FALSE);
317    }
318 
319    util_blitter_save_vertex_buffer_slot(ctx->blitter, ctx->vertex_buffer);
320    util_blitter_save_vertex_elements(ctx->blitter, (void *)ctx->velems);
321    util_blitter_save_vertex_shader(ctx->blitter, (void *)ctx->vs);
322    util_blitter_save_geometry_shader(ctx->blitter, (void*)ctx->gs);
323    util_blitter_save_tessctrl_shader(ctx->blitter, (void*)ctx->tcs);
324    util_blitter_save_tesseval_shader(ctx->blitter, (void*)ctx->tes);
325    util_blitter_save_so_targets(
326       ctx->blitter,
327       ctx->num_so_targets,
328       (struct pipe_stream_output_target **)ctx->so_targets);
329    util_blitter_save_rasterizer(ctx->blitter, (void *)ctx->rasterizer);
330    util_blitter_save_viewport(ctx->blitter, &ctx->viewports[0]);
331    util_blitter_save_scissor(ctx->blitter, &ctx->scissors[0]);
332    util_blitter_save_fragment_shader(ctx->blitter, ctx->fs);
333    util_blitter_save_blend(ctx->blitter, (void *)ctx->blend);
334    util_blitter_save_depth_stencil_alpha(ctx->blitter,
335                                          (void *)ctx->depth_stencil);
336    util_blitter_save_stencil_ref(ctx->blitter, &ctx->stencil_ref);
337    util_blitter_save_sample_mask(ctx->blitter, ctx->sample_mask);
338    util_blitter_save_framebuffer(ctx->blitter, &ctx->framebuffer);
339    util_blitter_save_fragment_sampler_states(
340       ctx->blitter,
341       ctx->num_samplers[PIPE_SHADER_FRAGMENT],
342       (void **)ctx->samplers[PIPE_SHADER_FRAGMENT]);
343    util_blitter_save_fragment_sampler_views(
344       ctx->blitter,
345       ctx->num_sampler_views[PIPE_SHADER_FRAGMENT],
346       ctx->sampler_views[PIPE_SHADER_FRAGMENT]);
347    util_blitter_save_render_condition(ctx->blitter,
348                                       ctx->render_cond_query,
349                                       ctx->render_cond_cond,
350                                       ctx->render_cond_mode);
351 
352    util_blitter_blit(ctx->blitter, &info);
353 
354    if (ctx->active_queries) {
355       ctx->api.pfnSwrEnableStatsFE(ctx->swrContext, TRUE);
356       ctx->api.pfnSwrEnableStatsBE(ctx->swrContext, TRUE);
357    }
358 }
359 
360 
361 static void
swr_destroy(struct pipe_context * pipe)362 swr_destroy(struct pipe_context *pipe)
363 {
364    struct swr_context *ctx = swr_context(pipe);
365    struct swr_screen *screen = swr_screen(pipe->screen);
366 
367    if (ctx->blitter)
368       util_blitter_destroy(ctx->blitter);
369 
370    for (unsigned i = 0; i < PIPE_MAX_COLOR_BUFS; i++) {
371       if (ctx->framebuffer.cbufs[i]) {
372          struct swr_resource *res = swr_resource(ctx->framebuffer.cbufs[i]->texture);
373          /* NULL curr_pipe, so we don't have a reference to a deleted pipe */
374          res->curr_pipe = NULL;
375          pipe_surface_reference(&ctx->framebuffer.cbufs[i], NULL);
376       }
377    }
378 
379    if (ctx->framebuffer.zsbuf) {
380       struct swr_resource *res = swr_resource(ctx->framebuffer.zsbuf->texture);
381       /* NULL curr_pipe, so we don't have a reference to a deleted pipe */
382       res->curr_pipe = NULL;
383       pipe_surface_reference(&ctx->framebuffer.zsbuf, NULL);
384    }
385 
386    for (unsigned i = 0; i < ARRAY_SIZE(ctx->sampler_views[0]); i++) {
387       pipe_sampler_view_reference(&ctx->sampler_views[PIPE_SHADER_FRAGMENT][i], NULL);
388    }
389 
390    for (unsigned i = 0; i < ARRAY_SIZE(ctx->sampler_views[0]); i++) {
391       pipe_sampler_view_reference(&ctx->sampler_views[PIPE_SHADER_VERTEX][i], NULL);
392    }
393 
394    if (ctx->pipe.stream_uploader)
395       u_upload_destroy(ctx->pipe.stream_uploader);
396 
397    /* Idle core after destroying buffer resources, but before deleting
398     * context.  Destroying resources has potentially called StoreTiles.*/
399    ctx->api.pfnSwrWaitForIdle(ctx->swrContext);
400 
401    if (ctx->swrContext)
402       ctx->api.pfnSwrDestroyContext(ctx->swrContext);
403 
404    delete ctx->blendJIT;
405 
406    swr_destroy_scratch_buffers(ctx);
407 
408 
409    /* Only update screen->pipe if current context is being destroyed */
410    assert(screen);
411    if (screen->pipe == pipe)
412       screen->pipe = NULL;
413 
414    AlignedFree(ctx);
415 }
416 
417 
418 static void
swr_render_condition(struct pipe_context * pipe,struct pipe_query * query,bool condition,enum pipe_render_cond_flag mode)419 swr_render_condition(struct pipe_context *pipe,
420                      struct pipe_query *query,
421                      bool condition,
422                      enum pipe_render_cond_flag mode)
423 {
424    struct swr_context *ctx = swr_context(pipe);
425 
426    ctx->render_cond_query = query;
427    ctx->render_cond_mode = mode;
428    ctx->render_cond_cond = condition;
429 }
430 
431 static void
swr_UpdateStats(HANDLE hPrivateContext,const SWR_STATS * pStats)432 swr_UpdateStats(HANDLE hPrivateContext, const SWR_STATS *pStats)
433 {
434    swr_draw_context *pDC = (swr_draw_context*)hPrivateContext;
435 
436    if (!pDC)
437       return;
438 
439    struct swr_query_result *pqr = pDC->pStats;
440 
441    SWR_STATS *pSwrStats = &pqr->core;
442 
443    pSwrStats->DepthPassCount += pStats->DepthPassCount;
444    pSwrStats->PsInvocations += pStats->PsInvocations;
445    pSwrStats->CsInvocations += pStats->CsInvocations;
446 }
447 
448 static void
swr_UpdateStatsFE(HANDLE hPrivateContext,const SWR_STATS_FE * pStats)449 swr_UpdateStatsFE(HANDLE hPrivateContext, const SWR_STATS_FE *pStats)
450 {
451    swr_draw_context *pDC = (swr_draw_context*)hPrivateContext;
452 
453    if (!pDC)
454       return;
455 
456    struct swr_query_result *pqr = pDC->pStats;
457 
458    SWR_STATS_FE *pSwrStats = &pqr->coreFE;
459    p_atomic_add(&pSwrStats->IaVertices, pStats->IaVertices);
460    p_atomic_add(&pSwrStats->IaPrimitives, pStats->IaPrimitives);
461    p_atomic_add(&pSwrStats->VsInvocations, pStats->VsInvocations);
462    p_atomic_add(&pSwrStats->HsInvocations, pStats->HsInvocations);
463    p_atomic_add(&pSwrStats->DsInvocations, pStats->DsInvocations);
464    p_atomic_add(&pSwrStats->GsInvocations, pStats->GsInvocations);
465    p_atomic_add(&pSwrStats->CInvocations, pStats->CInvocations);
466    p_atomic_add(&pSwrStats->CPrimitives, pStats->CPrimitives);
467    p_atomic_add(&pSwrStats->GsPrimitives, pStats->GsPrimitives);
468 
469    for (unsigned i = 0; i < 4; i++) {
470       p_atomic_add(&pSwrStats->SoPrimStorageNeeded[i],
471             pStats->SoPrimStorageNeeded[i]);
472       p_atomic_add(&pSwrStats->SoNumPrimsWritten[i],
473             pStats->SoNumPrimsWritten[i]);
474    }
475 }
476 
477 static void
swr_UpdateStreamOut(HANDLE hPrivateContext,uint64_t numPrims)478 swr_UpdateStreamOut(HANDLE hPrivateContext, uint64_t numPrims)
479 {
480    swr_draw_context *pDC = (swr_draw_context*)hPrivateContext;
481 
482    if (!pDC)
483       return;
484 
485    if (pDC->soPrims)
486        *pDC->soPrims += numPrims;
487 }
488 
489 struct pipe_context *
swr_create_context(struct pipe_screen * p_screen,void * priv,unsigned flags)490 swr_create_context(struct pipe_screen *p_screen, void *priv, unsigned flags)
491 {
492    struct swr_context *ctx = (struct swr_context *)
493       AlignedMalloc(sizeof(struct swr_context), KNOB_SIMD_BYTES);
494    memset((void*)ctx, 0, sizeof(struct swr_context));
495 
496    swr_screen(p_screen)->pfnSwrGetInterface(ctx->api);
497    swr_screen(p_screen)->pfnSwrGetTileInterface(ctx->tileApi);
498    ctx->swrDC.pAPI = &ctx->api;
499    ctx->swrDC.pTileAPI = &ctx->tileApi;
500 
501    ctx->blendJIT =
502       new std::unordered_map<BLEND_COMPILE_STATE, PFN_BLEND_JIT_FUNC>;
503 
504    ctx->max_draws_in_flight = KNOB_MAX_DRAWS_IN_FLIGHT;
505 
506    SWR_CREATECONTEXT_INFO createInfo {0};
507 
508    createInfo.privateStateSize = sizeof(swr_draw_context);
509    createInfo.pfnLoadTile = swr_LoadHotTile;
510    createInfo.pfnStoreTile = swr_StoreHotTile;
511    createInfo.pfnUpdateStats = swr_UpdateStats;
512    createInfo.pfnUpdateStatsFE = swr_UpdateStatsFE;
513    createInfo.pfnUpdateStreamOut = swr_UpdateStreamOut;
514    createInfo.pfnMakeGfxPtr = swr_MakeGfxPtr;
515 
516    SWR_THREADING_INFO threadingInfo {0};
517 
518    threadingInfo.MAX_WORKER_THREADS        = KNOB_MAX_WORKER_THREADS;
519    threadingInfo.MAX_NUMA_NODES            = KNOB_MAX_NUMA_NODES;
520    threadingInfo.MAX_CORES_PER_NUMA_NODE   = KNOB_MAX_CORES_PER_NUMA_NODE;
521    threadingInfo.MAX_THREADS_PER_CORE      = KNOB_MAX_THREADS_PER_CORE;
522    threadingInfo.SINGLE_THREADED           = KNOB_SINGLE_THREADED;
523 
524    // Use non-standard settings for KNL
525    if (swr_screen(p_screen)->is_knl)
526    {
527       if (nullptr == getenv("KNOB_MAX_THREADS_PER_CORE"))
528          threadingInfo.MAX_THREADS_PER_CORE  = 2;
529 
530       if (nullptr == getenv("KNOB_MAX_DRAWS_IN_FLIGHT"))
531       {
532          ctx->max_draws_in_flight = 2048;
533          createInfo.MAX_DRAWS_IN_FLIGHT = ctx->max_draws_in_flight;
534       }
535    }
536 
537    createInfo.pThreadInfo = &threadingInfo;
538 
539    ctx->swrContext = ctx->api.pfnSwrCreateContext(&createInfo);
540 
541    ctx->api.pfnSwrInit();
542 
543    if (ctx->swrContext == NULL)
544       goto fail;
545 
546    ctx->pipe.screen = p_screen;
547    ctx->pipe.destroy = swr_destroy;
548    ctx->pipe.priv = priv;
549    ctx->pipe.create_surface = swr_create_surface;
550    ctx->pipe.surface_destroy = swr_surface_destroy;
551    ctx->pipe.transfer_map = swr_transfer_map;
552    ctx->pipe.transfer_unmap = swr_transfer_unmap;
553    ctx->pipe.transfer_flush_region = swr_transfer_flush_region;
554 
555    ctx->pipe.buffer_subdata = u_default_buffer_subdata;
556    ctx->pipe.texture_subdata = u_default_texture_subdata;
557 
558    ctx->pipe.clear_texture = util_clear_texture;
559    ctx->pipe.resource_copy_region = swr_resource_copy;
560    ctx->pipe.render_condition = swr_render_condition;
561 
562    swr_state_init(&ctx->pipe);
563    swr_clear_init(&ctx->pipe);
564    swr_draw_init(&ctx->pipe);
565    swr_query_init(&ctx->pipe);
566 
567    ctx->pipe.stream_uploader = u_upload_create_default(&ctx->pipe);
568    if (!ctx->pipe.stream_uploader)
569       goto fail;
570    ctx->pipe.const_uploader = ctx->pipe.stream_uploader;
571 
572    ctx->pipe.blit = swr_blit;
573    ctx->blitter = util_blitter_create(&ctx->pipe);
574    if (!ctx->blitter)
575       goto fail;
576 
577    swr_init_scratch_buffers(ctx);
578 
579    return &ctx->pipe;
580 
581 fail:
582    /* Should really validate the init steps and fail gracefully */
583    swr_destroy(&ctx->pipe);
584    return NULL;
585 }
586