• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /**************************************************************************
2  *
3  * Copyright 2006 VMware, Inc.
4  * All Rights Reserved.
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a
7  * copy of this software and associated documentation files (the
8  * "Software"), to deal in the Software without restriction, including
9  * without limitation the rights to use, copy, modify, merge, publish,
10  * distribute, sub license, and/or sell copies of the Software, and to
11  * permit persons to whom the Software is furnished to do so, subject to
12  * the following conditions:
13  *
14  * The above copyright notice and this permission notice (including the
15  * next paragraph) shall be included in all copies or substantial portions
16  * of the Software.
17  *
18  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS
19  * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
20  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT.
21  * IN NO EVENT SHALL VMWARE AND/OR ITS SUPPLIERS BE LIABLE FOR
22  * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
23  * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
24  * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
25  *
26  **************************************************************************/
27 
28 #include <GL/gl.h>
29 #include <GL/internal/dri_interface.h>
30 
31 #include "intel_batchbuffer.h"
32 #include "intel_chipset.h"
33 #include "intel_context.h"
34 #include "intel_mipmap_tree.h"
35 #include "intel_regions.h"
36 #include "intel_tex_layout.h"
37 #include "intel_tex.h"
38 #include "intel_blit.h"
39 
40 #include "main/enums.h"
41 #include "main/formats.h"
42 #include "main/glformats.h"
43 #include "main/teximage.h"
44 
45 #define FILE_DEBUG_FLAG DEBUG_MIPTREE
46 
47 static GLenum
target_to_target(GLenum target)48 target_to_target(GLenum target)
49 {
50    switch (target) {
51    case GL_TEXTURE_CUBE_MAP_POSITIVE_X_ARB:
52    case GL_TEXTURE_CUBE_MAP_NEGATIVE_X_ARB:
53    case GL_TEXTURE_CUBE_MAP_POSITIVE_Y_ARB:
54    case GL_TEXTURE_CUBE_MAP_NEGATIVE_Y_ARB:
55    case GL_TEXTURE_CUBE_MAP_POSITIVE_Z_ARB:
56    case GL_TEXTURE_CUBE_MAP_NEGATIVE_Z_ARB:
57       return GL_TEXTURE_CUBE_MAP_ARB;
58    default:
59       return target;
60    }
61 }
62 
63 struct intel_mipmap_tree *
intel_miptree_create_layout(struct intel_context * intel,GLenum target,mesa_format format,GLuint first_level,GLuint last_level,GLuint width0,GLuint height0,GLuint depth0)64 intel_miptree_create_layout(struct intel_context *intel,
65                             GLenum target,
66                             mesa_format format,
67                             GLuint first_level,
68                             GLuint last_level,
69                             GLuint width0,
70                             GLuint height0,
71                             GLuint depth0)
72 {
73    struct intel_mipmap_tree *mt = calloc(sizeof(*mt), 1);
74    if (!mt)
75       return NULL;
76 
77    DBG("%s target %s format %s level %d..%d <-- %p\n", __func__,
78        _mesa_enum_to_string(target),
79        _mesa_get_format_name(format),
80        first_level, last_level, mt);
81 
82    mt->target = target_to_target(target);
83    mt->format = format;
84    mt->first_level = first_level;
85    mt->last_level = last_level;
86    mt->logical_width0 = width0;
87    mt->logical_height0 = height0;
88    mt->logical_depth0 = depth0;
89 
90    /* The cpp is bytes per (1, blockheight)-sized block for compressed
91     * textures.  This is why you'll see divides by blockheight all over
92     */
93    unsigned bw, bh;
94    _mesa_get_format_block_size(format, &bw, &bh);
95    assert(_mesa_get_format_bytes(mt->format) % bw == 0);
96    mt->cpp = _mesa_get_format_bytes(mt->format) / bw;
97 
98    mt->compressed = _mesa_is_format_compressed(format);
99    mt->refcount = 1;
100 
101    if (target == GL_TEXTURE_CUBE_MAP) {
102       assert(depth0 == 1);
103       depth0 = 6;
104    }
105 
106    mt->physical_width0 = width0;
107    mt->physical_height0 = height0;
108    mt->physical_depth0 = depth0;
109 
110    intel_get_texture_alignment_unit(intel, mt->format,
111 				    &mt->align_w, &mt->align_h);
112 
113    (void) intel;
114    if (intel->is_945)
115       i945_miptree_layout(mt);
116    else
117       i915_miptree_layout(mt);
118 
119    return mt;
120 }
121 
122 /**
123  * \brief Helper function for intel_miptree_create().
124  */
125 static uint32_t
intel_miptree_choose_tiling(struct intel_context * intel,mesa_format format,uint32_t width0,enum intel_miptree_tiling_mode requested,struct intel_mipmap_tree * mt)126 intel_miptree_choose_tiling(struct intel_context *intel,
127                             mesa_format format,
128                             uint32_t width0,
129                             enum intel_miptree_tiling_mode requested,
130                             struct intel_mipmap_tree *mt)
131 {
132    /* Some usages may want only one type of tiling, like depth miptrees (Y
133     * tiled), or temporary BOs for uploading data once (linear).
134     */
135    switch (requested) {
136    case INTEL_MIPTREE_TILING_ANY:
137       break;
138    case INTEL_MIPTREE_TILING_Y:
139       return I915_TILING_Y;
140    case INTEL_MIPTREE_TILING_NONE:
141       return I915_TILING_NONE;
142    }
143 
144    int minimum_pitch = mt->total_width * mt->cpp;
145 
146    /* If the width is much smaller than a tile, don't bother tiling. */
147    if (minimum_pitch < 64)
148       return I915_TILING_NONE;
149 
150    if (ALIGN(minimum_pitch, 512) >= 32768) {
151       perf_debug("%dx%d miptree too large to blit, falling back to untiled",
152                  mt->total_width, mt->total_height);
153       return I915_TILING_NONE;
154    }
155 
156    /* We don't have BLORP to handle Y-tiled blits, so use X-tiling. */
157    return I915_TILING_X;
158 }
159 
160 struct intel_mipmap_tree *
intel_miptree_create(struct intel_context * intel,GLenum target,mesa_format format,GLuint first_level,GLuint last_level,GLuint width0,GLuint height0,GLuint depth0,bool expect_accelerated_upload,enum intel_miptree_tiling_mode requested_tiling)161 intel_miptree_create(struct intel_context *intel,
162 		     GLenum target,
163 		     mesa_format format,
164 		     GLuint first_level,
165 		     GLuint last_level,
166 		     GLuint width0,
167 		     GLuint height0,
168 		     GLuint depth0,
169 		     bool expect_accelerated_upload,
170                      enum intel_miptree_tiling_mode requested_tiling)
171 {
172    struct intel_mipmap_tree *mt;
173    GLuint total_width, total_height;
174 
175 
176    mt = intel_miptree_create_layout(intel, target, format,
177 				      first_level, last_level, width0,
178 				      height0, depth0);
179    /*
180     * pitch == 0 || height == 0  indicates the null texture
181     */
182    if (!mt || !mt->total_width || !mt->total_height) {
183       intel_miptree_release(&mt);
184       return NULL;
185    }
186 
187    total_width = mt->total_width;
188    total_height = mt->total_height;
189 
190    uint32_t tiling = intel_miptree_choose_tiling(intel, format, width0,
191                                                  requested_tiling,
192                                                  mt);
193    bool y_or_x = tiling == (I915_TILING_Y | I915_TILING_X);
194 
195    mt->region = intel_region_alloc(intel->intelScreen,
196 				   y_or_x ? I915_TILING_Y : tiling,
197 				   mt->cpp,
198 				   total_width,
199 				   total_height,
200 				   expect_accelerated_upload);
201 
202    /* If the region is too large to fit in the aperture, we need to use the
203     * BLT engine to support it.  The BLT paths can't currently handle Y-tiling,
204     * so we need to fall back to X.
205     */
206    if (y_or_x && mt->region->bo->size >= intel->max_gtt_map_object_size) {
207       perf_debug("%dx%d miptree larger than aperture; falling back to X-tiled\n",
208                  mt->total_width, mt->total_height);
209       intel_region_release(&mt->region);
210 
211       mt->region = intel_region_alloc(intel->intelScreen,
212                                       I915_TILING_X,
213                                       mt->cpp,
214                                       total_width,
215                                       total_height,
216                                       expect_accelerated_upload);
217    }
218 
219    mt->offset = 0;
220 
221    if (!mt->region) {
222        intel_miptree_release(&mt);
223        return NULL;
224    }
225 
226    return mt;
227 }
228 
229 struct intel_mipmap_tree *
intel_miptree_create_for_bo(struct intel_context * intel,drm_intel_bo * bo,mesa_format format,uint32_t offset,uint32_t width,uint32_t height,int pitch,uint32_t tiling)230 intel_miptree_create_for_bo(struct intel_context *intel,
231                             drm_intel_bo *bo,
232                             mesa_format format,
233                             uint32_t offset,
234                             uint32_t width,
235                             uint32_t height,
236                             int pitch,
237                             uint32_t tiling)
238 {
239    struct intel_mipmap_tree *mt;
240 
241    struct intel_region *region = calloc(1, sizeof(*region));
242    if (!region)
243       return NULL;
244 
245    /* Nothing will be able to use this miptree with the BO if the offset isn't
246     * aligned.
247     */
248    if (tiling != I915_TILING_NONE)
249       assert(offset % 4096 == 0);
250 
251    /* miptrees can't handle negative pitch.  If you need flipping of images,
252     * that's outside of the scope of the mt.
253     */
254    assert(pitch >= 0);
255 
256    mt = intel_miptree_create_layout(intel, GL_TEXTURE_2D, format,
257                                     0, 0,
258                                     width, height, 1);
259    if (!mt) {
260       free(region);
261       return mt;
262    }
263 
264    region->cpp = mt->cpp;
265    region->width = width;
266    region->height = height;
267    region->pitch = pitch;
268    region->refcount = 1;
269    drm_intel_bo_reference(bo);
270    region->bo = bo;
271    region->tiling = tiling;
272 
273    mt->region = region;
274    mt->offset = offset;
275 
276    return mt;
277 }
278 
279 
280 /**
281  * For a singlesample DRI2 buffer, this simply wraps the given region with a miptree.
282  *
283  * For a multisample DRI2 buffer, this wraps the given region with
284  * a singlesample miptree, then creates a multisample miptree into which the
285  * singlesample miptree is embedded as a child.
286  */
287 struct intel_mipmap_tree*
intel_miptree_create_for_dri2_buffer(struct intel_context * intel,unsigned dri_attachment,mesa_format format,struct intel_region * region)288 intel_miptree_create_for_dri2_buffer(struct intel_context *intel,
289                                      unsigned dri_attachment,
290                                      mesa_format format,
291                                      struct intel_region *region)
292 {
293    struct intel_mipmap_tree *mt = NULL;
294 
295    /* Only the front and back buffers, which are color buffers, are shared
296     * through DRI2.
297     */
298    assert(dri_attachment == __DRI_BUFFER_BACK_LEFT ||
299           dri_attachment == __DRI_BUFFER_FRONT_LEFT ||
300           dri_attachment == __DRI_BUFFER_FAKE_FRONT_LEFT);
301    assert(_mesa_get_format_base_format(format) == GL_RGB ||
302           _mesa_get_format_base_format(format) == GL_RGBA);
303 
304    mt = intel_miptree_create_for_bo(intel,
305                                     region->bo,
306                                     format,
307                                     0,
308                                     region->width,
309                                     region->height,
310                                     region->pitch,
311                                     region->tiling);
312    if (!mt)
313       return NULL;
314    mt->region->name = region->name;
315 
316    return mt;
317 }
318 
319 /**
320  * For a singlesample image buffer, this simply wraps the given region with a miptree.
321  *
322  * For a multisample image buffer, this wraps the given region with
323  * a singlesample miptree, then creates a multisample miptree into which the
324  * singlesample miptree is embedded as a child.
325  */
326 struct intel_mipmap_tree*
intel_miptree_create_for_image_buffer(struct intel_context * intel,enum __DRIimageBufferMask buffer_type,mesa_format format,uint32_t num_samples,struct intel_region * region)327 intel_miptree_create_for_image_buffer(struct intel_context *intel,
328                                       enum __DRIimageBufferMask buffer_type,
329                                       mesa_format format,
330                                       uint32_t num_samples,
331                                       struct intel_region *region)
332 {
333    struct intel_mipmap_tree *mt = NULL;
334 
335    /* Only the front and back buffers, which are color buffers, are allocated
336     * through the image loader.
337     */
338    assert(_mesa_get_format_base_format(format) == GL_RGB ||
339           _mesa_get_format_base_format(format) == GL_RGBA);
340 
341    mt = intel_miptree_create_for_bo(intel,
342                                     region->bo,
343                                     format,
344                                     0,
345                                     region->width,
346                                     region->height,
347                                     region->pitch,
348                                     region->tiling);
349    return mt;
350 }
351 
352 struct intel_mipmap_tree*
intel_miptree_create_for_renderbuffer(struct intel_context * intel,mesa_format format,uint32_t width,uint32_t height)353 intel_miptree_create_for_renderbuffer(struct intel_context *intel,
354                                       mesa_format format,
355                                       uint32_t width,
356                                       uint32_t height)
357 {
358    uint32_t depth = 1;
359 
360    return intel_miptree_create(intel, GL_TEXTURE_2D, format, 0, 0,
361                                width, height, depth, true,
362                                INTEL_MIPTREE_TILING_ANY);
363 }
364 
365 void
intel_miptree_reference(struct intel_mipmap_tree ** dst,struct intel_mipmap_tree * src)366 intel_miptree_reference(struct intel_mipmap_tree **dst,
367                         struct intel_mipmap_tree *src)
368 {
369    if (*dst == src)
370       return;
371 
372    intel_miptree_release(dst);
373 
374    if (src) {
375       src->refcount++;
376       DBG("%s %p refcount now %d\n", __func__, src, src->refcount);
377    }
378 
379    *dst = src;
380 }
381 
382 
383 void
intel_miptree_release(struct intel_mipmap_tree ** mt)384 intel_miptree_release(struct intel_mipmap_tree **mt)
385 {
386    if (!*mt)
387       return;
388 
389    DBG("%s %p refcount will be %d\n", __func__, *mt, (*mt)->refcount - 1);
390    if (--(*mt)->refcount <= 0) {
391       GLuint i;
392 
393       DBG("%s deleting %p\n", __func__, *mt);
394 
395       intel_region_release(&((*mt)->region));
396 
397       for (i = 0; i < MAX_TEXTURE_LEVELS; i++) {
398 	 free((*mt)->level[i].slice);
399       }
400 
401       free(*mt);
402    }
403    *mt = NULL;
404 }
405 
406 void
intel_miptree_get_dimensions_for_image(struct gl_texture_image * image,int * width,int * height,int * depth)407 intel_miptree_get_dimensions_for_image(struct gl_texture_image *image,
408                                        int *width, int *height, int *depth)
409 {
410    switch (image->TexObject->Target) {
411    case GL_TEXTURE_1D_ARRAY:
412       *width = image->Width;
413       *height = 1;
414       *depth = image->Height;
415       break;
416    default:
417       *width = image->Width;
418       *height = image->Height;
419       *depth = image->Depth;
420       break;
421    }
422 }
423 
424 /**
425  * Can the image be pulled into a unified mipmap tree?  This mirrors
426  * the completeness test in a lot of ways.
427  *
428  * Not sure whether I want to pass gl_texture_image here.
429  */
430 bool
intel_miptree_match_image(struct intel_mipmap_tree * mt,struct gl_texture_image * image)431 intel_miptree_match_image(struct intel_mipmap_tree *mt,
432                           struct gl_texture_image *image)
433 {
434    struct intel_texture_image *intelImage = intel_texture_image(image);
435    GLuint level = intelImage->base.Base.Level;
436    int width, height, depth;
437 
438    /* glTexImage* choose the texture object based on the target passed in, and
439     * objects can't change targets over their lifetimes, so this should be
440     * true.
441     */
442    assert(target_to_target(image->TexObject->Target) == mt->target);
443 
444    mesa_format mt_format = mt->format;
445 
446    if (image->TexFormat != mt_format)
447       return false;
448 
449    intel_miptree_get_dimensions_for_image(image, &width, &height, &depth);
450 
451    if (mt->target == GL_TEXTURE_CUBE_MAP)
452       depth = 6;
453 
454    /* Test image dimensions against the base level image adjusted for
455     * minification.  This will also catch images not present in the
456     * tree, changed targets, etc.
457     */
458    if (mt->target == GL_TEXTURE_2D_MULTISAMPLE ||
459          mt->target == GL_TEXTURE_2D_MULTISAMPLE_ARRAY) {
460       /* nonzero level here is always bogus */
461       assert(level == 0);
462 
463       if (width != mt->logical_width0 ||
464             height != mt->logical_height0 ||
465             depth != mt->logical_depth0) {
466          return false;
467       }
468    }
469    else {
470       /* all normal textures, renderbuffers, etc */
471       if (width != mt->level[level].width ||
472           height != mt->level[level].height ||
473           depth != mt->level[level].depth) {
474          return false;
475       }
476    }
477 
478    return true;
479 }
480 
481 
482 void
intel_miptree_set_level_info(struct intel_mipmap_tree * mt,GLuint level,GLuint x,GLuint y,GLuint w,GLuint h,GLuint d)483 intel_miptree_set_level_info(struct intel_mipmap_tree *mt,
484 			     GLuint level,
485 			     GLuint x, GLuint y,
486 			     GLuint w, GLuint h, GLuint d)
487 {
488    mt->level[level].width = w;
489    mt->level[level].height = h;
490    mt->level[level].depth = d;
491    mt->level[level].level_x = x;
492    mt->level[level].level_y = y;
493 
494    DBG("%s level %d size: %d,%d,%d offset %d,%d\n", __func__,
495        level, w, h, d, x, y);
496 
497    assert(mt->level[level].slice == NULL);
498 
499    mt->level[level].slice = calloc(d, sizeof(*mt->level[0].slice));
500    mt->level[level].slice[0].x_offset = mt->level[level].level_x;
501    mt->level[level].slice[0].y_offset = mt->level[level].level_y;
502 }
503 
504 
505 void
intel_miptree_set_image_offset(struct intel_mipmap_tree * mt,GLuint level,GLuint img,GLuint x,GLuint y)506 intel_miptree_set_image_offset(struct intel_mipmap_tree *mt,
507 			       GLuint level, GLuint img,
508 			       GLuint x, GLuint y)
509 {
510    if (img == 0 && level == 0)
511       assert(x == 0 && y == 0);
512 
513    assert(img < mt->level[level].depth);
514 
515    mt->level[level].slice[img].x_offset = mt->level[level].level_x + x;
516    mt->level[level].slice[img].y_offset = mt->level[level].level_y + y;
517 
518    DBG("%s level %d img %d pos %d,%d\n",
519        __func__, level, img,
520        mt->level[level].slice[img].x_offset,
521        mt->level[level].slice[img].y_offset);
522 }
523 
524 void
intel_miptree_get_image_offset(struct intel_mipmap_tree * mt,GLuint level,GLuint slice,GLuint * x,GLuint * y)525 intel_miptree_get_image_offset(struct intel_mipmap_tree *mt,
526 			       GLuint level, GLuint slice,
527 			       GLuint *x, GLuint *y)
528 {
529    assert(slice < mt->level[level].depth);
530 
531    *x = mt->level[level].slice[slice].x_offset;
532    *y = mt->level[level].slice[slice].y_offset;
533 }
534 
535 /**
536  * Rendering with tiled buffers requires that the base address of the buffer
537  * be aligned to a page boundary.  For renderbuffers, and sometimes with
538  * textures, we may want the surface to point at a texture image level that
539  * isn't at a page boundary.
540  *
541  * This function returns an appropriately-aligned base offset
542  * according to the tiling restrictions, plus any required x/y offset
543  * from there.
544  */
545 uint32_t
intel_miptree_get_tile_offsets(struct intel_mipmap_tree * mt,GLuint level,GLuint slice,uint32_t * tile_x,uint32_t * tile_y)546 intel_miptree_get_tile_offsets(struct intel_mipmap_tree *mt,
547                                GLuint level, GLuint slice,
548                                uint32_t *tile_x,
549                                uint32_t *tile_y)
550 {
551    struct intel_region *region = mt->region;
552    uint32_t x, y;
553    uint32_t mask_x, mask_y;
554 
555    intel_region_get_tile_masks(region, &mask_x, &mask_y, false);
556    intel_miptree_get_image_offset(mt, level, slice, &x, &y);
557 
558    *tile_x = x & mask_x;
559    *tile_y = y & mask_y;
560 
561    return intel_region_get_aligned_offset(region, x & ~mask_x, y & ~mask_y,
562                                           false);
563 }
564 
565 static void
intel_miptree_copy_slice_sw(struct intel_context * intel,struct intel_mipmap_tree * dst_mt,struct intel_mipmap_tree * src_mt,int level,int slice,int width,int height)566 intel_miptree_copy_slice_sw(struct intel_context *intel,
567                             struct intel_mipmap_tree *dst_mt,
568                             struct intel_mipmap_tree *src_mt,
569                             int level,
570                             int slice,
571                             int width,
572                             int height)
573 {
574    void *src, *dst;
575    int src_stride, dst_stride;
576    int cpp = dst_mt->cpp;
577 
578    intel_miptree_map(intel, src_mt,
579                      level, slice,
580                      0, 0,
581                      width, height,
582                      GL_MAP_READ_BIT,
583                      &src, &src_stride);
584 
585    intel_miptree_map(intel, dst_mt,
586                      level, slice,
587                      0, 0,
588                      width, height,
589                      GL_MAP_WRITE_BIT | GL_MAP_INVALIDATE_RANGE_BIT,
590                      &dst, &dst_stride);
591 
592    DBG("sw blit %s mt %p %p/%d -> %s mt %p %p/%d (%dx%d)\n",
593        _mesa_get_format_name(src_mt->format),
594        src_mt, src, src_stride,
595        _mesa_get_format_name(dst_mt->format),
596        dst_mt, dst, dst_stride,
597        width, height);
598 
599    int row_size = cpp * width;
600    if (src_stride == row_size &&
601        dst_stride == row_size) {
602       memcpy(dst, src, row_size * height);
603    } else {
604       for (int i = 0; i < height; i++) {
605          memcpy(dst, src, row_size);
606          dst += dst_stride;
607          src += src_stride;
608       }
609    }
610 
611    intel_miptree_unmap(intel, dst_mt, level, slice);
612    intel_miptree_unmap(intel, src_mt, level, slice);
613 }
614 
615 static void
intel_miptree_copy_slice(struct intel_context * intel,struct intel_mipmap_tree * dst_mt,struct intel_mipmap_tree * src_mt,int level,int face,int depth)616 intel_miptree_copy_slice(struct intel_context *intel,
617 			 struct intel_mipmap_tree *dst_mt,
618 			 struct intel_mipmap_tree *src_mt,
619 			 int level,
620 			 int face,
621 			 int depth)
622 
623 {
624    mesa_format format = src_mt->format;
625    uint32_t width = src_mt->level[level].width;
626    uint32_t height = src_mt->level[level].height;
627    int slice;
628 
629    if (face > 0)
630       slice = face;
631    else
632       slice = depth;
633 
634    assert(depth < src_mt->level[level].depth);
635    assert(src_mt->format == dst_mt->format);
636 
637    if (dst_mt->compressed) {
638       height = ALIGN(height, dst_mt->align_h) / dst_mt->align_h;
639       width = ALIGN(width, dst_mt->align_w);
640    }
641 
642    uint32_t dst_x, dst_y, src_x, src_y;
643    intel_miptree_get_image_offset(dst_mt, level, slice, &dst_x, &dst_y);
644    intel_miptree_get_image_offset(src_mt, level, slice, &src_x, &src_y);
645 
646    DBG("validate blit mt %s %p %d,%d/%d -> mt %s %p %d,%d/%d (%dx%d)\n",
647        _mesa_get_format_name(src_mt->format),
648        src_mt, src_x, src_y, src_mt->region->pitch,
649        _mesa_get_format_name(dst_mt->format),
650        dst_mt, dst_x, dst_y, dst_mt->region->pitch,
651        width, height);
652 
653    if (!intel_miptree_blit(intel,
654                            src_mt, level, slice, 0, 0, false,
655                            dst_mt, level, slice, 0, 0, false,
656                            width, height, GL_COPY)) {
657       perf_debug("miptree validate blit for %s failed\n",
658                  _mesa_get_format_name(format));
659 
660       intel_miptree_copy_slice_sw(intel, dst_mt, src_mt, level, slice,
661                                   width, height);
662    }
663 }
664 
665 /**
666  * Copies the image's current data to the given miptree, and associates that
667  * miptree with the image.
668  *
669  * If \c invalidate is true, then the actual image data does not need to be
670  * copied, but the image still needs to be associated to the new miptree (this
671  * is set to true if we're about to clear the image).
672  */
673 void
intel_miptree_copy_teximage(struct intel_context * intel,struct intel_texture_image * intelImage,struct intel_mipmap_tree * dst_mt,bool invalidate)674 intel_miptree_copy_teximage(struct intel_context *intel,
675 			    struct intel_texture_image *intelImage,
676 			    struct intel_mipmap_tree *dst_mt,
677                             bool invalidate)
678 {
679    struct intel_mipmap_tree *src_mt = intelImage->mt;
680    struct intel_texture_object *intel_obj =
681       intel_texture_object(intelImage->base.Base.TexObject);
682    int level = intelImage->base.Base.Level;
683    int face = intelImage->base.Base.Face;
684    GLuint depth = intelImage->base.Base.Depth;
685 
686    if (!invalidate) {
687       for (int slice = 0; slice < depth; slice++) {
688          intel_miptree_copy_slice(intel, dst_mt, src_mt, level, face, slice);
689       }
690    }
691 
692    intel_miptree_reference(&intelImage->mt, dst_mt);
693    intel_obj->needs_validate = true;
694 }
695 
696 void *
intel_miptree_map_raw(struct intel_context * intel,struct intel_mipmap_tree * mt)697 intel_miptree_map_raw(struct intel_context *intel, struct intel_mipmap_tree *mt)
698 {
699    drm_intel_bo *bo = mt->region->bo;
700 
701    if (unlikely(INTEL_DEBUG & DEBUG_PERF)) {
702       if (drm_intel_bo_busy(bo)) {
703          perf_debug("Mapping a busy BO, causing a stall on the GPU.\n");
704       }
705    }
706 
707    intel_flush(&intel->ctx);
708 
709    if (mt->region->tiling != I915_TILING_NONE)
710       drm_intel_gem_bo_map_gtt(bo);
711    else
712       drm_intel_bo_map(bo, true);
713 
714    return bo->virtual;
715 }
716 
717 void
intel_miptree_unmap_raw(struct intel_mipmap_tree * mt)718 intel_miptree_unmap_raw(struct intel_mipmap_tree *mt)
719 {
720    drm_intel_bo_unmap(mt->region->bo);
721 }
722 
723 static void
intel_miptree_map_gtt(struct intel_context * intel,struct intel_mipmap_tree * mt,struct intel_miptree_map * map,unsigned int level,unsigned int slice)724 intel_miptree_map_gtt(struct intel_context *intel,
725 		      struct intel_mipmap_tree *mt,
726 		      struct intel_miptree_map *map,
727 		      unsigned int level, unsigned int slice)
728 {
729    unsigned int bw, bh;
730    void *base;
731    unsigned int image_x, image_y;
732    int x = map->x;
733    int y = map->y;
734 
735    /* For compressed formats, the stride is the number of bytes per
736     * row of blocks.  intel_miptree_get_image_offset() already does
737     * the divide.
738     */
739    _mesa_get_format_block_size(mt->format, &bw, &bh);
740    assert(y % bh == 0);
741    y /= bh;
742 
743    base = intel_miptree_map_raw(intel, mt) + mt->offset;
744 
745    if (base == NULL)
746       map->ptr = NULL;
747    else {
748       /* Note that in the case of cube maps, the caller must have passed the
749        * slice number referencing the face.
750       */
751       intel_miptree_get_image_offset(mt, level, slice, &image_x, &image_y);
752       x += image_x;
753       y += image_y;
754 
755       map->stride = mt->region->pitch;
756       map->ptr = base + y * map->stride + x * mt->cpp;
757    }
758 
759    DBG("%s: %d,%d %dx%d from mt %p (%s) %d,%d = %p/%d\n", __func__,
760        map->x, map->y, map->w, map->h,
761        mt, _mesa_get_format_name(mt->format),
762        x, y, map->ptr, map->stride);
763 }
764 
765 static void
intel_miptree_unmap_gtt(struct intel_mipmap_tree * mt)766 intel_miptree_unmap_gtt(struct intel_mipmap_tree *mt)
767 {
768    intel_miptree_unmap_raw(mt);
769 }
770 
771 static void
intel_miptree_map_blit(struct intel_context * intel,struct intel_mipmap_tree * mt,struct intel_miptree_map * map,unsigned int level,unsigned int slice)772 intel_miptree_map_blit(struct intel_context *intel,
773 		       struct intel_mipmap_tree *mt,
774 		       struct intel_miptree_map *map,
775 		       unsigned int level, unsigned int slice)
776 {
777    map->mt = intel_miptree_create(intel, GL_TEXTURE_2D, mt->format,
778                                   0, 0,
779                                   map->w, map->h, 1,
780                                   false,
781                                   INTEL_MIPTREE_TILING_NONE);
782    if (!map->mt) {
783       fprintf(stderr, "Failed to allocate blit temporary\n");
784       goto fail;
785    }
786    map->stride = map->mt->region->pitch;
787 
788    if (!intel_miptree_blit(intel,
789                            mt, level, slice,
790                            map->x, map->y, false,
791                            map->mt, 0, 0,
792                            0, 0, false,
793                            map->w, map->h, GL_COPY)) {
794       fprintf(stderr, "Failed to blit\n");
795       goto fail;
796    }
797 
798    intel_batchbuffer_flush(intel);
799    map->ptr = intel_miptree_map_raw(intel, map->mt);
800 
801    DBG("%s: %d,%d %dx%d from mt %p (%s) %d,%d = %p/%d\n", __func__,
802        map->x, map->y, map->w, map->h,
803        mt, _mesa_get_format_name(mt->format),
804        level, slice, map->ptr, map->stride);
805 
806    return;
807 
808 fail:
809    intel_miptree_release(&map->mt);
810    map->ptr = NULL;
811    map->stride = 0;
812 }
813 
814 static void
intel_miptree_unmap_blit(struct intel_context * intel,struct intel_mipmap_tree * mt,struct intel_miptree_map * map,unsigned int level,unsigned int slice)815 intel_miptree_unmap_blit(struct intel_context *intel,
816 			 struct intel_mipmap_tree *mt,
817 			 struct intel_miptree_map *map,
818 			 unsigned int level,
819 			 unsigned int slice)
820 {
821    struct gl_context *ctx = &intel->ctx;
822 
823    intel_miptree_unmap_raw(map->mt);
824 
825    if (map->mode & GL_MAP_WRITE_BIT) {
826       bool ok = intel_miptree_blit(intel,
827                                    map->mt, 0, 0,
828                                    0, 0, false,
829                                    mt, level, slice,
830                                    map->x, map->y, false,
831                                    map->w, map->h, GL_COPY);
832       WARN_ONCE(!ok, "Failed to blit from linear temporary mapping");
833    }
834 
835    intel_miptree_release(&map->mt);
836 }
837 
838 /**
839  * Create and attach a map to the miptree at (level, slice). Return the
840  * attached map.
841  */
842 static struct intel_miptree_map*
intel_miptree_attach_map(struct intel_mipmap_tree * mt,unsigned int level,unsigned int slice,unsigned int x,unsigned int y,unsigned int w,unsigned int h,GLbitfield mode)843 intel_miptree_attach_map(struct intel_mipmap_tree *mt,
844                          unsigned int level,
845                          unsigned int slice,
846                          unsigned int x,
847                          unsigned int y,
848                          unsigned int w,
849                          unsigned int h,
850                          GLbitfield mode)
851 {
852    struct intel_miptree_map *map = calloc(1, sizeof(*map));
853 
854    if (!map)
855       return NULL;
856 
857    assert(mt->level[level].slice[slice].map == NULL);
858    mt->level[level].slice[slice].map = map;
859 
860    map->mode = mode;
861    map->x = x;
862    map->y = y;
863    map->w = w;
864    map->h = h;
865 
866    return map;
867 }
868 
869 /**
870  * Release the map at (level, slice).
871  */
872 static void
intel_miptree_release_map(struct intel_mipmap_tree * mt,unsigned int level,unsigned int slice)873 intel_miptree_release_map(struct intel_mipmap_tree *mt,
874                          unsigned int level,
875                          unsigned int slice)
876 {
877    struct intel_miptree_map **map;
878 
879    map = &mt->level[level].slice[slice].map;
880    free(*map);
881    *map = NULL;
882 }
883 
884 void
intel_miptree_map(struct intel_context * intel,struct intel_mipmap_tree * mt,unsigned int level,unsigned int slice,unsigned int x,unsigned int y,unsigned int w,unsigned int h,GLbitfield mode,void ** out_ptr,int * out_stride)885 intel_miptree_map(struct intel_context *intel,
886                   struct intel_mipmap_tree *mt,
887                   unsigned int level,
888                   unsigned int slice,
889                   unsigned int x,
890                   unsigned int y,
891                   unsigned int w,
892                   unsigned int h,
893                   GLbitfield mode,
894                   void **out_ptr,
895                   int *out_stride)
896 {
897    struct intel_miptree_map *map;
898 
899    map = intel_miptree_attach_map(mt, level, slice, x, y, w, h, mode);
900    if (!map){
901       *out_ptr = NULL;
902       *out_stride = 0;
903       return;
904    }
905 
906    /* See intel_miptree_blit() for details on the 32k pitch limit. */
907    if (mt->region->tiling != I915_TILING_NONE &&
908        mt->region->bo->size >= intel->max_gtt_map_object_size) {
909       assert(mt->region->pitch < 32768);
910       intel_miptree_map_blit(intel, mt, map, level, slice);
911    } else {
912       intel_miptree_map_gtt(intel, mt, map, level, slice);
913    }
914 
915    *out_ptr = map->ptr;
916    *out_stride = map->stride;
917 
918    if (map->ptr == NULL)
919       intel_miptree_release_map(mt, level, slice);
920 }
921 
922 void
intel_miptree_unmap(struct intel_context * intel,struct intel_mipmap_tree * mt,unsigned int level,unsigned int slice)923 intel_miptree_unmap(struct intel_context *intel,
924                     struct intel_mipmap_tree *mt,
925                     unsigned int level,
926                     unsigned int slice)
927 {
928    struct intel_miptree_map *map = mt->level[level].slice[slice].map;
929 
930    if (!map)
931       return;
932 
933    DBG("%s: mt %p (%s) level %d slice %d\n", __func__,
934        mt, _mesa_get_format_name(mt->format), level, slice);
935 
936    if (map->mt) {
937       intel_miptree_unmap_blit(intel, mt, map, level, slice);
938    } else {
939       intel_miptree_unmap_gtt(mt);
940    }
941 
942    intel_miptree_release_map(mt, level, slice);
943 }
944