• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright 2018 Collabora Ltd.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * on the rights to use, copy, modify, merge, publish, distribute, sub
8  * license, and/or sell copies of the Software, and to permit persons to whom
9  * the Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice (including the next
12  * paragraph) shall be included in all copies or substantial portions of the
13  * Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17  * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
18  * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
19  * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
20  * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
21  * USE OR OTHER DEALINGS IN THE SOFTWARE.
22  */
23 
24 #include "zink_resource.h"
25 
26 #include "zink_batch.h"
27 #include "zink_clear.h"
28 #include "zink_context.h"
29 #include "zink_fence.h"
30 #include "zink_format.h"
31 #include "zink_program.h"
32 #include "zink_screen.h"
33 #include "zink_kopper.h"
34 
35 #ifdef VK_USE_PLATFORM_METAL_EXT
36 #include "QuartzCore/CAMetalLayer.h"
37 #endif
38 
39 #include "vk_format.h"
40 #include "util/u_blitter.h"
41 #include "util/u_debug.h"
42 #include "util/format/u_format.h"
43 #include "util/u_transfer_helper.h"
44 #include "util/u_inlines.h"
45 #include "util/u_memory.h"
46 #include "util/u_upload_mgr.h"
47 #include "util/os_file.h"
48 #include "frontend/winsys_handle.h"
49 
50 #if !defined(__APPLE__)
51 #define ZINK_USE_DMABUF
52 #endif
53 
54 #if defined(ZINK_USE_DMABUF) && !defined(_WIN32)
55 #include "drm-uapi/drm_fourcc.h"
56 #else
57 /* these won't actually be used */
58 #define DRM_FORMAT_MOD_INVALID 0
59 #define DRM_FORMAT_MOD_LINEAR 0
60 #endif
61 
62 #if defined(__APPLE__)
63 // Source of MVK_VERSION
64 #include "MoltenVK/vk_mvk_moltenvk.h"
65 #endif
66 
67 #define ZINK_EXTERNAL_MEMORY_HANDLE 999
68 
69 
70 
71 struct zink_debug_mem_entry {
72    uint32_t count;
73    uint64_t size;
74    const char *name;
75 };
76 
77 static const char *
zink_debug_mem_add(struct zink_screen * screen,uint64_t size,const char * name)78 zink_debug_mem_add(struct zink_screen *screen, uint64_t size, const char *name)
79 {
80    assert(name);
81 
82    simple_mtx_lock(&screen->debug_mem_lock);
83    struct hash_entry *entry = _mesa_hash_table_search(screen->debug_mem_sizes, name);
84    struct zink_debug_mem_entry *debug_bos;
85 
86    if (!entry) {
87       debug_bos = calloc(1, sizeof(struct zink_debug_mem_entry));
88       debug_bos->name = strdup(name);
89       _mesa_hash_table_insert(screen->debug_mem_sizes, debug_bos->name, debug_bos);
90    } else {
91       debug_bos = (struct zink_debug_mem_entry *) entry->data;
92    }
93 
94    debug_bos->count++;
95    debug_bos->size += align(size, 4096);
96    simple_mtx_unlock(&screen->debug_mem_lock);
97 
98    return debug_bos->name;
99 }
100 
101 static void
zink_debug_mem_del(struct zink_screen * screen,struct zink_bo * bo)102 zink_debug_mem_del(struct zink_screen *screen, struct zink_bo *bo)
103 {
104    simple_mtx_lock(&screen->debug_mem_lock);
105    struct hash_entry *entry = _mesa_hash_table_search(screen->debug_mem_sizes, bo->name);
106    /* If we're finishing the BO, it should have been added already */
107    assert(entry);
108 
109    struct zink_debug_mem_entry *debug_bos = entry->data;
110    debug_bos->count--;
111    debug_bos->size -= align(zink_bo_get_size(bo), 4096);
112    if (!debug_bos->count) {
113       _mesa_hash_table_remove(screen->debug_mem_sizes, entry);
114       free((void*)debug_bos->name);
115       free(debug_bos);
116    }
117    simple_mtx_unlock(&screen->debug_mem_lock);
118 }
119 
120 static int
debug_bos_count_compare(const void * in_a,const void * in_b)121 debug_bos_count_compare(const void *in_a, const void *in_b)
122 {
123    struct zink_debug_mem_entry *a = *(struct zink_debug_mem_entry **)in_a;
124    struct zink_debug_mem_entry *b = *(struct zink_debug_mem_entry **)in_b;
125    return a->count - b->count;
126 }
127 
128 void
zink_debug_mem_print_stats(struct zink_screen * screen)129 zink_debug_mem_print_stats(struct zink_screen *screen)
130 {
131    simple_mtx_lock(&screen->debug_mem_lock);
132 
133    /* Put the HT's sizes data in an array so we can sort by number of allocations. */
134    struct util_dynarray dyn;
135    util_dynarray_init(&dyn, NULL);
136 
137    uint32_t size = 0;
138    uint32_t count = 0;
139    hash_table_foreach(screen->debug_mem_sizes, entry)
140    {
141       struct zink_debug_mem_entry *debug_bos = entry->data;
142       util_dynarray_append(&dyn, struct zink_debug_mem_entry *, debug_bos);
143       size += debug_bos->size / 1024;
144       count += debug_bos->count;
145    }
146 
147    qsort(dyn.data,
148          util_dynarray_num_elements(&dyn, struct zink_debug_mem_entry *),
149          sizeof(struct zink_debug_mem_entryos_entry *), debug_bos_count_compare);
150 
151    util_dynarray_foreach(&dyn, struct zink_debug_mem_entry *, entryp)
152    {
153       struct zink_debug_mem_entry *debug_bos = *entryp;
154       mesa_logi("%30s: %4d bos, %lld kb\n", debug_bos->name, debug_bos->count,
155                 (long long) (debug_bos->size / 1024));
156    }
157 
158    mesa_logi("submitted %d bos (%d MB)\n", count, DIV_ROUND_UP(size, 1024));
159 
160    util_dynarray_fini(&dyn);
161 
162    simple_mtx_unlock(&screen->debug_mem_lock);
163 }
164 
165 static bool
equals_ivci(const void * a,const void * b)166 equals_ivci(const void *a, const void *b)
167 {
168    const uint8_t *pa = a;
169    const uint8_t *pb = b;
170    size_t offset = offsetof(VkImageViewCreateInfo, flags);
171    return memcmp(pa + offset, pb + offset, sizeof(VkImageViewCreateInfo) - offset) == 0;
172 }
173 
174 static bool
equals_bvci(const void * a,const void * b)175 equals_bvci(const void *a, const void *b)
176 {
177    const uint8_t *pa = a;
178    const uint8_t *pb = b;
179    size_t offset = offsetof(VkBufferViewCreateInfo, flags);
180    return memcmp(pa + offset, pb + offset, sizeof(VkBufferViewCreateInfo) - offset) == 0;
181 }
182 
183 static void
184 zink_transfer_flush_region(struct pipe_context *pctx,
185                            struct pipe_transfer *ptrans,
186                            const struct pipe_box *box);
187 
188 void
debug_describe_zink_resource_object(char * buf,const struct zink_resource_object * ptr)189 debug_describe_zink_resource_object(char *buf, const struct zink_resource_object *ptr)
190 {
191    sprintf(buf, "zink_resource_object");
192 }
193 
194 void
zink_destroy_resource_object(struct zink_screen * screen,struct zink_resource_object * obj)195 zink_destroy_resource_object(struct zink_screen *screen, struct zink_resource_object *obj)
196 {
197    if (obj->is_buffer) {
198       while (util_dynarray_contains(&obj->views, VkBufferView))
199          VKSCR(DestroyBufferView)(screen->dev, util_dynarray_pop(&obj->views, VkBufferView), NULL);
200    } else {
201       while (util_dynarray_contains(&obj->views, VkImageView))
202          VKSCR(DestroyImageView)(screen->dev, util_dynarray_pop(&obj->views, VkImageView), NULL);
203    }
204    if (!obj->dt && zink_debug & ZINK_DEBUG_MEM)
205       zink_debug_mem_del(screen, obj->bo);
206    util_dynarray_fini(&obj->views);
207    for (unsigned i = 0; i < ARRAY_SIZE(obj->copies); i++)
208       util_dynarray_fini(&obj->copies[i]);
209    if (obj->is_buffer) {
210       VKSCR(DestroyBuffer)(screen->dev, obj->buffer, NULL);
211       VKSCR(DestroyBuffer)(screen->dev, obj->storage_buffer, NULL);
212    } else if (obj->dt) {
213       zink_kopper_displaytarget_destroy(screen, obj->dt);
214    } else if (!obj->is_aux) {
215       VKSCR(DestroyImage)(screen->dev, obj->image, NULL);
216    } else {
217 #if defined(ZINK_USE_DMABUF) && !defined(_WIN32)
218       close(obj->handle);
219 #endif
220    }
221 
222    simple_mtx_destroy(&obj->view_lock);
223    if (obj->dt) {
224       FREE(obj->bo); //this is a dummy struct
225    } else
226       zink_bo_unref(screen, obj->bo);
227    FREE(obj);
228 }
229 
230 static void
zink_resource_destroy(struct pipe_screen * pscreen,struct pipe_resource * pres)231 zink_resource_destroy(struct pipe_screen *pscreen,
232                       struct pipe_resource *pres)
233 {
234    struct zink_screen *screen = zink_screen(pscreen);
235    struct zink_resource *res = zink_resource(pres);
236    if (pres->target == PIPE_BUFFER) {
237       util_range_destroy(&res->valid_buffer_range);
238       util_idalloc_mt_free(&screen->buffer_ids, res->base.buffer_id_unique);
239       assert(!_mesa_hash_table_num_entries(&res->bufferview_cache));
240       simple_mtx_destroy(&res->bufferview_mtx);
241       ralloc_free(res->bufferview_cache.table);
242    } else {
243       assert(!_mesa_hash_table_num_entries(&res->surface_cache));
244       simple_mtx_destroy(&res->surface_mtx);
245       ralloc_free(res->surface_cache.table);
246    }
247    /* no need to do anything for the caches, these objects own the resource lifetimes */
248 
249    zink_resource_object_reference(screen, &res->obj, NULL);
250    threaded_resource_deinit(pres);
251    FREE_CL(res);
252 }
253 
254 static VkImageAspectFlags
aspect_from_format(enum pipe_format fmt)255 aspect_from_format(enum pipe_format fmt)
256 {
257    if (util_format_is_depth_or_stencil(fmt)) {
258       VkImageAspectFlags aspect = 0;
259       const struct util_format_description *desc = util_format_description(fmt);
260       if (util_format_has_depth(desc))
261          aspect |= VK_IMAGE_ASPECT_DEPTH_BIT;
262       if (util_format_has_stencil(desc))
263          aspect |= VK_IMAGE_ASPECT_STENCIL_BIT;
264       return aspect;
265    } else
266      return VK_IMAGE_ASPECT_COLOR_BIT;
267 }
268 
269 static VkBufferCreateInfo
create_bci(struct zink_screen * screen,const struct pipe_resource * templ,unsigned bind)270 create_bci(struct zink_screen *screen, const struct pipe_resource *templ, unsigned bind)
271 {
272    VkBufferCreateInfo bci;
273    bci.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
274    bci.pNext = NULL;
275    bci.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
276    bci.queueFamilyIndexCount = 0;
277    bci.pQueueFamilyIndices = NULL;
278    bci.size = templ->width0;
279    bci.flags = 0;
280    assert(bci.size > 0);
281 
282    if (bind & ZINK_BIND_DESCRIPTOR) {
283       /* gallium sizes are all uint32_t, while the total size of this buffer may exceed that limit */
284       bci.usage = 0;
285       if (bind & ZINK_BIND_SAMPLER_DESCRIPTOR)
286          bci.usage |= VK_BUFFER_USAGE_SAMPLER_DESCRIPTOR_BUFFER_BIT_EXT;
287       if (bind & ZINK_BIND_RESOURCE_DESCRIPTOR)
288          bci.usage |= VK_BUFFER_USAGE_RESOURCE_DESCRIPTOR_BUFFER_BIT_EXT;
289    } else {
290       bci.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT |
291                   VK_BUFFER_USAGE_TRANSFER_DST_BIT |
292                   VK_BUFFER_USAGE_STORAGE_BUFFER_BIT;
293 
294       bci.usage |= VK_BUFFER_USAGE_UNIFORM_TEXEL_BUFFER_BIT |
295                   VK_BUFFER_USAGE_INDIRECT_BUFFER_BIT |
296                   VK_BUFFER_USAGE_VERTEX_BUFFER_BIT |
297                   VK_BUFFER_USAGE_INDEX_BUFFER_BIT |
298                   VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT |
299                   VK_BUFFER_USAGE_TRANSFORM_FEEDBACK_BUFFER_BIT_EXT |
300                   VK_BUFFER_USAGE_TRANSFORM_FEEDBACK_COUNTER_BUFFER_BIT_EXT;
301    }
302    if (screen->info.have_KHR_buffer_device_address)
303       bci.usage |= VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT;
304 
305    if (bind & PIPE_BIND_SHADER_IMAGE)
306       bci.usage |= VK_BUFFER_USAGE_STORAGE_TEXEL_BUFFER_BIT;
307 
308    if (bind & PIPE_BIND_QUERY_BUFFER)
309       bci.usage |= VK_BUFFER_USAGE_CONDITIONAL_RENDERING_BIT_EXT;
310 
311    if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE)
312       bci.flags |= VK_BUFFER_CREATE_SPARSE_BINDING_BIT | VK_BUFFER_CREATE_SPARSE_RESIDENCY_BIT;
313    return bci;
314 }
315 
316 typedef enum {
317    USAGE_FAIL_NONE,
318    USAGE_FAIL_ERROR,
319    USAGE_FAIL_SUBOPTIMAL,
320 } usage_fail;
321 
322 static usage_fail
check_ici(struct zink_screen * screen,VkImageCreateInfo * ici,uint64_t modifier)323 check_ici(struct zink_screen *screen, VkImageCreateInfo *ici, uint64_t modifier)
324 {
325    VkImageFormatProperties image_props;
326    VkResult ret;
327    bool optimalDeviceAccess = true;
328    assert(modifier == DRM_FORMAT_MOD_INVALID ||
329           (VKSCR(GetPhysicalDeviceImageFormatProperties2) && screen->info.have_EXT_image_drm_format_modifier));
330    if (VKSCR(GetPhysicalDeviceImageFormatProperties2)) {
331       VkImageFormatProperties2 props2;
332       props2.sType = VK_STRUCTURE_TYPE_IMAGE_FORMAT_PROPERTIES_2;
333       props2.pNext = NULL;
334       VkSamplerYcbcrConversionImageFormatProperties ycbcr_props;
335       ycbcr_props.sType = VK_STRUCTURE_TYPE_SAMPLER_YCBCR_CONVERSION_IMAGE_FORMAT_PROPERTIES;
336       ycbcr_props.pNext = NULL;
337       if (screen->info.have_KHR_sampler_ycbcr_conversion)
338          props2.pNext = &ycbcr_props;
339       VkHostImageCopyDevicePerformanceQueryEXT hic = {
340          VK_STRUCTURE_TYPE_HOST_IMAGE_COPY_DEVICE_PERFORMANCE_QUERY_EXT,
341          props2.pNext,
342       };
343       if (screen->info.have_EXT_host_image_copy && ici->usage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT)
344          props2.pNext = &hic;
345       VkPhysicalDeviceImageFormatInfo2 info;
346       info.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_IMAGE_FORMAT_INFO_2;
347       /* possibly VkImageFormatListCreateInfo */
348       info.pNext = ici->pNext;
349       info.format = ici->format;
350       info.type = ici->imageType;
351       info.tiling = ici->tiling;
352       info.usage = ici->usage;
353       info.flags = ici->flags;
354 
355       VkPhysicalDeviceImageDrmFormatModifierInfoEXT mod_info;
356       if (modifier != DRM_FORMAT_MOD_INVALID) {
357          mod_info.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_IMAGE_DRM_FORMAT_MODIFIER_INFO_EXT;
358          mod_info.pNext = info.pNext;
359          mod_info.drmFormatModifier = modifier;
360          mod_info.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
361          mod_info.queueFamilyIndexCount = 0;
362          mod_info.pQueueFamilyIndices = NULL;
363          info.pNext = &mod_info;
364       }
365 
366       ret = VKSCR(GetPhysicalDeviceImageFormatProperties2)(screen->pdev, &info, &props2);
367       /* this is using VK_IMAGE_CREATE_EXTENDED_USAGE_BIT and can't be validated */
368       if (vk_format_aspects(ici->format) & VK_IMAGE_ASPECT_PLANE_1_BIT)
369          ret = VK_SUCCESS;
370       image_props = props2.imageFormatProperties;
371       if (screen->info.have_EXT_host_image_copy && ici->usage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT)
372          optimalDeviceAccess = hic.optimalDeviceAccess;
373    } else
374       ret = VKSCR(GetPhysicalDeviceImageFormatProperties)(screen->pdev, ici->format, ici->imageType,
375                                                    ici->tiling, ici->usage, ici->flags, &image_props);
376    if (ret != VK_SUCCESS)
377       return USAGE_FAIL_ERROR;
378    if (ici->extent.depth > image_props.maxExtent.depth ||
379        ici->extent.height > image_props.maxExtent.height ||
380        ici->extent.width > image_props.maxExtent.width)
381       return USAGE_FAIL_ERROR;
382    if (ici->mipLevels > image_props.maxMipLevels)
383       return USAGE_FAIL_ERROR;
384    if (ici->arrayLayers > image_props.maxArrayLayers)
385       return USAGE_FAIL_ERROR;
386    if (!(ici->samples & image_props.sampleCounts))
387       return USAGE_FAIL_ERROR;
388    if (!optimalDeviceAccess)
389       return USAGE_FAIL_SUBOPTIMAL;
390    return USAGE_FAIL_NONE;
391 }
392 
393 static VkImageUsageFlags
get_image_usage_for_feats(struct zink_screen * screen,VkFormatFeatureFlags2 feats,const struct pipe_resource * templ,unsigned bind,bool * need_extended)394 get_image_usage_for_feats(struct zink_screen *screen, VkFormatFeatureFlags2 feats, const struct pipe_resource *templ, unsigned bind, bool *need_extended)
395 {
396    VkImageUsageFlags usage = 0;
397    bool is_planar = util_format_get_num_planes(templ->format) > 1;
398    *need_extended = false;
399 
400    if (bind & ZINK_BIND_TRANSIENT)
401       usage |= VK_IMAGE_USAGE_TRANSIENT_ATTACHMENT_BIT;
402    else {
403       /* sadly, gallium doesn't let us know if it'll ever need this, so we have to assume */
404       if (is_planar || (feats & VK_FORMAT_FEATURE_TRANSFER_SRC_BIT))
405          usage |= VK_IMAGE_USAGE_TRANSFER_SRC_BIT;
406       if (is_planar || (feats & VK_FORMAT_FEATURE_TRANSFER_DST_BIT))
407          usage |= VK_IMAGE_USAGE_TRANSFER_DST_BIT;
408       if (feats & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT)
409          usage |= VK_IMAGE_USAGE_SAMPLED_BIT;
410 
411       if ((is_planar || (feats & VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT)) && (bind & PIPE_BIND_SHADER_IMAGE)) {
412          assert(templ->nr_samples <= 1 || screen->info.feats.features.shaderStorageImageMultisample);
413          usage |= VK_IMAGE_USAGE_STORAGE_BIT;
414       }
415    }
416 
417    if (bind & PIPE_BIND_RENDER_TARGET) {
418       if (feats & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT) {
419          usage |= VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
420          if (!(bind & ZINK_BIND_TRANSIENT) && (bind & (PIPE_BIND_LINEAR | PIPE_BIND_SHARED)) != (PIPE_BIND_LINEAR | PIPE_BIND_SHARED))
421             usage |= VK_IMAGE_USAGE_INPUT_ATTACHMENT_BIT;
422          if (!(bind & ZINK_BIND_TRANSIENT) && screen->info.have_EXT_attachment_feedback_loop_layout)
423             usage |= VK_IMAGE_USAGE_ATTACHMENT_FEEDBACK_LOOP_BIT_EXT;
424       } else {
425          /* trust that gallium isn't going to give us anything wild */
426          *need_extended = true;
427          return 0;
428       }
429    } else if ((bind & PIPE_BIND_SAMPLER_VIEW) && !util_format_is_depth_or_stencil(templ->format)) {
430       if (!(feats & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT)) {
431          /* ensure we can u_blitter this later */
432          *need_extended = true;
433          return 0;
434       }
435       usage |= VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
436    }
437 
438    if (bind & PIPE_BIND_DEPTH_STENCIL) {
439       if (feats & VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT)
440          usage |= VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT;
441       else
442          return 0;
443       if (screen->info.have_EXT_attachment_feedback_loop_layout && !(bind & ZINK_BIND_TRANSIENT))
444          usage |= VK_IMAGE_USAGE_ATTACHMENT_FEEDBACK_LOOP_BIT_EXT;
445    /* this is unlikely to occur and has been included for completeness */
446    } else if (bind & PIPE_BIND_SAMPLER_VIEW && !(usage & VK_IMAGE_USAGE_TRANSFER_DST_BIT)) {
447       if (feats & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT)
448          usage |= VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
449       else
450          return 0;
451    }
452 
453    if (bind & PIPE_BIND_STREAM_OUTPUT)
454       usage |= VK_IMAGE_USAGE_INPUT_ATTACHMENT_BIT;
455 
456    if (screen->info.have_EXT_host_image_copy && feats & VK_FORMAT_FEATURE_2_HOST_IMAGE_TRANSFER_BIT_EXT)
457       usage |= VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT;
458 
459    return usage;
460 }
461 
462 static VkFormatFeatureFlags
find_modifier_feats(const struct zink_modifier_prop * prop,uint64_t modifier,uint64_t * mod)463 find_modifier_feats(const struct zink_modifier_prop *prop, uint64_t modifier, uint64_t *mod)
464 {
465    for (unsigned j = 0; j < prop->drmFormatModifierCount; j++) {
466       if (prop->pDrmFormatModifierProperties[j].drmFormatModifier == modifier) {
467          *mod = modifier;
468          return prop->pDrmFormatModifierProperties[j].drmFormatModifierTilingFeatures;
469       }
470    }
471    return 0;
472 }
473 
474 /* check HIC optimalness */
475 static bool
suboptimal_check_ici(struct zink_screen * screen,VkImageCreateInfo * ici,uint64_t * mod)476 suboptimal_check_ici(struct zink_screen *screen, VkImageCreateInfo *ici, uint64_t *mod)
477 {
478    usage_fail fail = check_ici(screen, ici, *mod);
479    if (!fail)
480       return true;
481    if (fail == USAGE_FAIL_SUBOPTIMAL) {
482       ici->usage &= ~VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT;
483       fail = check_ici(screen, ici, *mod);
484       if (!fail)
485          return true;
486    }
487    return false;
488 }
489 
490 /* If the driver can't do mutable with this ICI, then try again after removing mutable (and
491  * thus also the list of formats we might might mutate to)
492  */
493 static bool
double_check_ici(struct zink_screen * screen,VkImageCreateInfo * ici,VkImageUsageFlags usage,uint64_t * mod)494 double_check_ici(struct zink_screen *screen, VkImageCreateInfo *ici, VkImageUsageFlags usage, uint64_t *mod)
495 {
496    if (!usage)
497       return false;
498 
499    ici->usage = usage;
500 
501    if (suboptimal_check_ici(screen, ici, mod))
502       return true;
503    usage_fail fail = check_ici(screen, ici, *mod);
504    if (!fail)
505       return true;
506    if (fail == USAGE_FAIL_SUBOPTIMAL) {
507       ici->usage &= ~VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT;
508       fail = check_ici(screen, ici, *mod);
509       if (!fail)
510          return true;
511    }
512    const void *pNext = ici->pNext;
513    if (pNext) {
514       VkBaseOutStructure *prev = NULL;
515       VkBaseOutStructure *fmt_list = NULL;
516       vk_foreach_struct(strct, (void*)ici->pNext) {
517          if (strct->sType == VK_STRUCTURE_TYPE_IMAGE_FORMAT_LIST_CREATE_INFO) {
518             fmt_list = strct;
519             if (prev) {
520                prev->pNext = strct->pNext;
521             } else {
522                ici->pNext = strct->pNext;
523             }
524             fmt_list->pNext = NULL;
525             break;
526          }
527          prev = strct;
528       }
529       ici->flags &= ~VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
530       if (suboptimal_check_ici(screen, ici, mod))
531          return true;
532       fmt_list->pNext = (void*)ici->pNext;
533       ici->pNext = fmt_list;
534       ici->flags |= VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
535    }
536    return false;
537 }
538 
539 static VkImageUsageFlags
get_image_usage(struct zink_screen * screen,VkImageCreateInfo * ici,const struct pipe_resource * templ,unsigned bind,unsigned modifiers_count,uint64_t * modifiers,uint64_t * mod)540 get_image_usage(struct zink_screen *screen, VkImageCreateInfo *ici, const struct pipe_resource *templ, unsigned bind, unsigned modifiers_count, uint64_t *modifiers, uint64_t *mod)
541 {
542    VkImageTiling tiling = ici->tiling;
543    bool need_extended = false;
544    *mod = DRM_FORMAT_MOD_INVALID;
545    if (modifiers_count) {
546       bool have_linear = false;
547       const struct zink_modifier_prop *prop = &screen->modifier_props[templ->format];
548       assert(tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT);
549       bool found = false;
550       uint64_t good_mod = 0;
551       VkImageUsageFlags good_usage = 0;
552       for (unsigned i = 0; i < modifiers_count; i++) {
553          if (modifiers[i] == DRM_FORMAT_MOD_LINEAR) {
554             have_linear = true;
555             if (!screen->info.have_EXT_image_drm_format_modifier)
556                break;
557             continue;
558          }
559          VkFormatFeatureFlags feats = find_modifier_feats(prop, modifiers[i], mod);
560          if (feats) {
561             VkImageUsageFlags usage = get_image_usage_for_feats(screen, feats, templ, bind, &need_extended);
562             assert(!need_extended);
563             if (double_check_ici(screen, ici, usage, mod)) {
564                if (!found) {
565                   found = true;
566                   good_mod = modifiers[i];
567                   good_usage = usage;
568                }
569             } else {
570                modifiers[i] = DRM_FORMAT_MOD_LINEAR;
571             }
572          }
573       }
574       if (found) {
575          *mod = good_mod;
576          return good_usage;
577       }
578       /* only try linear if no other options available */
579       if (have_linear) {
580          VkFormatFeatureFlags feats = find_modifier_feats(prop, DRM_FORMAT_MOD_LINEAR, mod);
581          if (feats) {
582             VkImageUsageFlags usage = get_image_usage_for_feats(screen, feats, templ, bind, &need_extended);
583             assert(!need_extended);
584             if (double_check_ici(screen, ici, usage, mod))
585                return usage;
586          }
587       }
588    } else {
589       struct zink_format_props props = screen->format_props[templ->format];
590       VkFormatFeatureFlags2 feats = tiling == VK_IMAGE_TILING_LINEAR ? props.linearTilingFeatures : props.optimalTilingFeatures;
591       if (ici->flags & VK_IMAGE_CREATE_EXTENDED_USAGE_BIT)
592          feats = UINT32_MAX;
593       VkImageUsageFlags usage = get_image_usage_for_feats(screen, feats, templ, bind, &need_extended);
594       if (need_extended) {
595          ici->flags |= VK_IMAGE_CREATE_EXTENDED_USAGE_BIT | VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
596          feats = UINT32_MAX;
597          usage = get_image_usage_for_feats(screen, feats, templ, bind, &need_extended);
598       }
599       if (double_check_ici(screen, ici, usage, mod))
600          return usage;
601       if (util_format_is_depth_or_stencil(templ->format)) {
602          if (!(templ->bind & PIPE_BIND_DEPTH_STENCIL)) {
603             usage &= ~VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT;
604             if (double_check_ici(screen, ici, usage, mod))
605                return usage;
606          }
607       } else if (!(templ->bind & PIPE_BIND_RENDER_TARGET)) {
608          usage &= ~VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
609          if (double_check_ici(screen, ici, usage, mod))
610             return usage;
611       }
612    }
613    *mod = DRM_FORMAT_MOD_INVALID;
614    return 0;
615 }
616 
617 static uint64_t
eval_ici(struct zink_screen * screen,VkImageCreateInfo * ici,const struct pipe_resource * templ,unsigned bind,unsigned modifiers_count,uint64_t * modifiers,bool * success)618 eval_ici(struct zink_screen *screen, VkImageCreateInfo *ici, const struct pipe_resource *templ, unsigned bind, unsigned modifiers_count, uint64_t *modifiers, bool *success)
619 {
620    /* sampleCounts will be set to VK_SAMPLE_COUNT_1_BIT if at least one of the following conditions is true:
621     * - flags contains VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT
622     *
623     * 44.1.1. Supported Sample Counts
624     */
625    bool want_cube = ici->samples == 1 &&
626                     (templ->target == PIPE_TEXTURE_CUBE ||
627                     templ->target == PIPE_TEXTURE_CUBE_ARRAY ||
628                     (templ->target == PIPE_TEXTURE_2D_ARRAY && ici->extent.width == ici->extent.height && ici->arrayLayers >= 6));
629 
630    if (ici->tiling != VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT)
631       modifiers_count = 0;
632 
633    bool first = true;
634    bool tried[2] = {0};
635    uint64_t mod = DRM_FORMAT_MOD_INVALID;
636 retry:
637    while (!ici->usage) {
638       if (!first) {
639          switch (ici->tiling) {
640          case VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT:
641             ici->tiling = VK_IMAGE_TILING_OPTIMAL;
642             modifiers_count = 0;
643             break;
644          case VK_IMAGE_TILING_OPTIMAL:
645             ici->tiling = VK_IMAGE_TILING_LINEAR;
646             break;
647          case VK_IMAGE_TILING_LINEAR:
648             if (bind & PIPE_BIND_LINEAR) {
649                *success = false;
650                return DRM_FORMAT_MOD_INVALID;
651             }
652             ici->tiling = VK_IMAGE_TILING_OPTIMAL;
653             break;
654          default:
655             unreachable("unhandled tiling mode");
656          }
657          if (tried[ici->tiling]) {
658             if (ici->flags & VK_IMAGE_CREATE_EXTENDED_USAGE_BIT) {
659                *success = false;
660                return DRM_FORMAT_MOD_INVALID;
661             }
662             ici->flags |= VK_IMAGE_CREATE_EXTENDED_USAGE_BIT | VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
663             tried[0] = false;
664             tried[1] = false;
665             first = true;
666             goto retry;
667          }
668       }
669       ici->usage = get_image_usage(screen, ici, templ, bind, modifiers_count, modifiers, &mod);
670       first = false;
671       if (ici->tiling != VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT)
672          tried[ici->tiling] = true;
673    }
674    if (want_cube) {
675       ici->flags |= VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT;
676       if (get_image_usage(screen, ici, templ, bind, modifiers_count, modifiers, &mod) != ici->usage)
677          ici->flags &= ~VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT;
678    }
679 
680    *success = true;
681    return mod;
682 }
683 
684 static void
init_ici(struct zink_screen * screen,VkImageCreateInfo * ici,const struct pipe_resource * templ,unsigned bind,unsigned modifiers_count)685 init_ici(struct zink_screen *screen, VkImageCreateInfo *ici, const struct pipe_resource *templ, unsigned bind, unsigned modifiers_count)
686 {
687    ici->sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO;
688    /* pNext may already be set */
689    if (util_format_get_num_planes(templ->format) > 1)
690       ici->flags = VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT | VK_IMAGE_CREATE_EXTENDED_USAGE_BIT;
691    else if (bind & ZINK_BIND_MUTABLE)
692       ici->flags = VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
693    else
694       ici->flags = 0;
695    if (ici->flags & VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT)
696       /* unset VkImageFormatListCreateInfo if mutable */
697       ici->pNext = NULL;
698    else if (ici->pNext)
699       /* add mutable if VkImageFormatListCreateInfo */
700       ici->flags |= VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
701    ici->usage = 0;
702    ici->queueFamilyIndexCount = 0;
703    ici->pQueueFamilyIndices = NULL;
704 
705    /* assume we're going to be doing some CompressedTexSubImage */
706    if (util_format_is_compressed(templ->format) && (ici->flags & VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT) &&
707        !vk_find_struct_const(ici->pNext, IMAGE_FORMAT_LIST_CREATE_INFO))
708       ici->flags |= VK_IMAGE_CREATE_BLOCK_TEXEL_VIEW_COMPATIBLE_BIT;
709 
710    if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE)
711       ici->flags |= VK_IMAGE_CREATE_SPARSE_BINDING_BIT | VK_IMAGE_CREATE_SPARSE_RESIDENCY_BIT;
712 
713    bool need_2D = false;
714    switch (templ->target) {
715    case PIPE_TEXTURE_1D:
716    case PIPE_TEXTURE_1D_ARRAY:
717       if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE)
718          need_2D |= screen->need_2D_sparse;
719       if (util_format_is_depth_or_stencil(templ->format))
720          need_2D |= screen->need_2D_zs;
721       ici->imageType = need_2D ? VK_IMAGE_TYPE_2D : VK_IMAGE_TYPE_1D;
722       break;
723 
724    case PIPE_TEXTURE_CUBE:
725    case PIPE_TEXTURE_CUBE_ARRAY:
726    case PIPE_TEXTURE_2D:
727    case PIPE_TEXTURE_2D_ARRAY:
728    case PIPE_TEXTURE_RECT:
729       ici->imageType = VK_IMAGE_TYPE_2D;
730       break;
731 
732    case PIPE_TEXTURE_3D:
733       ici->imageType = VK_IMAGE_TYPE_3D;
734       if (!(templ->flags & PIPE_RESOURCE_FLAG_SPARSE))
735          ici->flags |= VK_IMAGE_CREATE_2D_ARRAY_COMPATIBLE_BIT;
736       if (screen->info.have_EXT_image_2d_view_of_3d)
737          ici->flags |= VK_IMAGE_CREATE_2D_VIEW_COMPATIBLE_BIT_EXT;
738       break;
739 
740    case PIPE_BUFFER:
741       unreachable("PIPE_BUFFER should already be handled");
742 
743    default:
744       unreachable("Unknown target");
745    }
746 
747    if (screen->info.have_EXT_sample_locations &&
748        bind & PIPE_BIND_DEPTH_STENCIL &&
749        util_format_has_depth(util_format_description(templ->format)))
750       ici->flags |= VK_IMAGE_CREATE_SAMPLE_LOCATIONS_COMPATIBLE_DEPTH_BIT_EXT;
751 
752    ici->format = zink_get_format(screen, templ->format);
753    ici->extent.width = templ->width0;
754    ici->extent.height = templ->height0;
755    ici->extent.depth = templ->depth0;
756    ici->mipLevels = templ->last_level + 1;
757    ici->arrayLayers = MAX2(templ->array_size, 1);
758    ici->samples = templ->nr_samples ? templ->nr_samples : VK_SAMPLE_COUNT_1_BIT;
759    ici->tiling = screen->info.have_EXT_image_drm_format_modifier && modifiers_count ?
760                  VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT :
761                  bind & (PIPE_BIND_LINEAR | ZINK_BIND_DMABUF) ? VK_IMAGE_TILING_LINEAR : VK_IMAGE_TILING_OPTIMAL;
762    /* XXX: does this have perf implications anywhere? hopefully not */
763    if (ici->samples == VK_SAMPLE_COUNT_1_BIT &&
764       screen->info.have_EXT_multisampled_render_to_single_sampled &&
765       ici->tiling != VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT)
766       ici->flags |= VK_IMAGE_CREATE_MULTISAMPLED_RENDER_TO_SINGLE_SAMPLED_BIT_EXT;
767    ici->sharingMode = VK_SHARING_MODE_EXCLUSIVE;
768    ici->initialLayout = VK_IMAGE_LAYOUT_UNDEFINED;
769 
770    if (templ->target == PIPE_TEXTURE_CUBE)
771       ici->arrayLayers *= 6;
772 }
773 
774 static inline bool
create_sampler_conversion(VkImageCreateInfo ici,struct zink_screen * screen,struct zink_resource_object * obj)775 create_sampler_conversion(VkImageCreateInfo ici, struct zink_screen *screen,
776                           struct zink_resource_object *obj)
777 {
778    if (obj->vkfeats & VK_FORMAT_FEATURE_DISJOINT_BIT)
779       ici.flags |= VK_IMAGE_CREATE_DISJOINT_BIT;
780    VkSamplerYcbcrConversionCreateInfo sycci = {0};
781    sycci.sType = VK_STRUCTURE_TYPE_SAMPLER_YCBCR_CONVERSION_CREATE_INFO;
782    sycci.pNext = NULL;
783    sycci.format = VK_FORMAT_G8_B8R8_2PLANE_420_UNORM;
784    sycci.ycbcrModel = VK_SAMPLER_YCBCR_MODEL_CONVERSION_YCBCR_709;
785    sycci.ycbcrRange = VK_SAMPLER_YCBCR_RANGE_ITU_FULL;
786    sycci.components.r = VK_COMPONENT_SWIZZLE_IDENTITY;
787    sycci.components.g = VK_COMPONENT_SWIZZLE_IDENTITY;
788    sycci.components.b = VK_COMPONENT_SWIZZLE_IDENTITY;
789    sycci.components.a = VK_COMPONENT_SWIZZLE_IDENTITY;
790    if (!obj->vkfeats || (obj->vkfeats & VK_FORMAT_FEATURE_COSITED_CHROMA_SAMPLES_BIT)) {
791       sycci.xChromaOffset = VK_CHROMA_LOCATION_COSITED_EVEN;
792       sycci.yChromaOffset = VK_CHROMA_LOCATION_COSITED_EVEN;
793    } else {
794       assert(obj->vkfeats & VK_FORMAT_FEATURE_MIDPOINT_CHROMA_SAMPLES_BIT);
795       sycci.xChromaOffset = VK_CHROMA_LOCATION_MIDPOINT;
796       sycci.yChromaOffset = VK_CHROMA_LOCATION_MIDPOINT;
797    }
798    sycci.chromaFilter = VK_FILTER_LINEAR;
799    sycci.forceExplicitReconstruction = VK_FALSE;
800    VkResult res = VKSCR(CreateSamplerYcbcrConversion)(screen->dev, &sycci, NULL, &obj->sampler_conversion);
801    if (res != VK_SUCCESS) {
802       mesa_loge("ZINK: vkCreateSamplerYcbcrConversion failed");
803       return false;
804    }
805    return true;
806 }
807 
808 static const VkImageAspectFlags plane_aspects[] = {
809    VK_IMAGE_ASPECT_PLANE_0_BIT,
810    VK_IMAGE_ASPECT_PLANE_1_BIT,
811    VK_IMAGE_ASPECT_PLANE_2_BIT,
812 };
813 
814 static inline bool
get_image_memory_requirement(struct zink_screen * screen,struct zink_resource_object * obj,unsigned num_planes,VkMemoryRequirements * reqs)815 get_image_memory_requirement(struct zink_screen *screen, struct zink_resource_object *obj,
816                              unsigned num_planes, VkMemoryRequirements *reqs)
817 {
818    bool need_dedicated = false;
819    if (VKSCR(GetImageMemoryRequirements2)) {
820       VkMemoryRequirements2 req2;
821       req2.sType = VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2;
822       VkImageMemoryRequirementsInfo2 info2;
823       info2.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_REQUIREMENTS_INFO_2;
824       info2.pNext = NULL;
825       info2.image = obj->image;
826       VkMemoryDedicatedRequirements ded;
827       ded.sType = VK_STRUCTURE_TYPE_MEMORY_DEDICATED_REQUIREMENTS;
828       ded.pNext = NULL;
829       req2.pNext = &ded;
830       VkImagePlaneMemoryRequirementsInfo plane;
831       plane.sType = VK_STRUCTURE_TYPE_IMAGE_PLANE_MEMORY_REQUIREMENTS_INFO;
832       plane.pNext = NULL;
833       if (num_planes > 1)
834          info2.pNext = &plane;
835       unsigned offset = 0;
836       for (unsigned i = 0; i < num_planes; i++) {
837          assert(i < ARRAY_SIZE(plane_aspects));
838          plane.planeAspect = plane_aspects[i];
839          VKSCR(GetImageMemoryRequirements2)(screen->dev, &info2, &req2);
840          if (!i)
841             reqs->alignment = req2.memoryRequirements.alignment;
842          obj->plane_offsets[i] = offset;
843          offset += req2.memoryRequirements.size;
844          reqs->size += req2.memoryRequirements.size;
845          reqs->memoryTypeBits |= req2.memoryRequirements.memoryTypeBits;
846          need_dedicated |= ded.prefersDedicatedAllocation || ded.requiresDedicatedAllocation;
847       }
848    } else {
849       VKSCR(GetImageMemoryRequirements)(screen->dev, obj->image, reqs);
850    }
851    return need_dedicated;
852 }
853 
854 static inline VkFormatFeatureFlags
get_format_feature_flags(VkImageCreateInfo ici,struct zink_screen * screen,const struct pipe_resource * templ)855 get_format_feature_flags(VkImageCreateInfo ici, struct zink_screen *screen, const struct pipe_resource *templ)
856 {
857    VkFormatFeatureFlags feats = 0;
858    switch (ici.tiling) {
859    case VK_IMAGE_TILING_LINEAR:
860       feats = screen->format_props[templ->format].linearTilingFeatures;
861       break;
862    case VK_IMAGE_TILING_OPTIMAL:
863       feats = screen->format_props[templ->format].optimalTilingFeatures;
864       break;
865    case VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT:
866       feats = VK_FORMAT_FEATURE_FLAG_BITS_MAX_ENUM;
867       /*
868          If is tiling then VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT, the value of
869          imageCreateFormatFeatures is found by calling vkGetPhysicalDeviceFormatProperties2
870          with VkImageFormatProperties::format equal to VkImageCreateInfo::format and with
871          VkDrmFormatModifierPropertiesListEXT chained into VkImageFormatProperties2; by
872          collecting all members of the returned array
873          VkDrmFormatModifierPropertiesListEXT::pDrmFormatModifierProperties
874          whose drmFormatModifier belongs to imageCreateDrmFormatModifiers; and by taking the bitwise
875          intersection, over the collected array members, of drmFormatModifierTilingFeatures.
876          (The resultant imageCreateFormatFeatures may be empty).
877          * -Chapter 12. Resource Creation
878        */
879       for (unsigned i = 0; i < screen->modifier_props[templ->format].drmFormatModifierCount; i++)
880          feats &= screen->modifier_props[templ->format].pDrmFormatModifierProperties[i].drmFormatModifierTilingFeatures;
881       break;
882    default:
883          unreachable("unknown tiling");
884    }
885    return feats;
886 }
887 
888 #if !defined(_WIN32)
889    #define ZINK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_BIT VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_FD_BIT
890 #else
891    #define ZINK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_BIT VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_WIN32_BIT
892 #endif
893 
894 
895 struct mem_alloc_info {
896    struct winsys_handle *whandle;
897    VkMemoryPropertyFlags flags;
898    enum zink_alloc_flag aflags;
899    bool need_dedicated;
900    bool shared;
901    const void *user_mem;
902    VkExternalMemoryHandleTypeFlags external;
903    VkExternalMemoryHandleTypeFlags export_types;
904 };
905 
906 static inline bool
get_export_flags(struct zink_screen * screen,const struct pipe_resource * templ,struct mem_alloc_info * alloc_info)907 get_export_flags(struct zink_screen *screen, const struct pipe_resource *templ, struct mem_alloc_info *alloc_info)
908 {
909    bool needs_export = (templ->bind & (ZINK_BIND_VIDEO | ZINK_BIND_DMABUF)) != 0;
910    if (alloc_info->whandle) {
911       if (alloc_info->whandle->type == WINSYS_HANDLE_TYPE_FD ||
912           alloc_info->whandle->type == ZINK_EXTERNAL_MEMORY_HANDLE)
913          needs_export |= true;
914       else
915          unreachable("unknown handle type");
916    }
917    if (needs_export) {
918       if (alloc_info->whandle && alloc_info->whandle->type == ZINK_EXTERNAL_MEMORY_HANDLE) {
919          alloc_info->external = ZINK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_BIT;
920       } else if (screen->info.have_EXT_external_memory_dma_buf) {
921          alloc_info->external = VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
922          alloc_info->export_types |= VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
923       } else {
924          return false;
925       }
926    }
927 
928    /* we may export WINSYS_HANDLE_TYPE_FD handle which is dma-buf */
929    if (templ->bind & PIPE_BIND_SHARED && screen->info.have_EXT_external_memory_dma_buf)
930       alloc_info->export_types |= VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
931    return true;
932 }
933 
934 enum resource_object_create_result {
935    roc_success,
936    roc_success_early_return,
937    roc_fail_and_free_object,
938    roc_fail_and_cleanup_object,
939    roc_fail_and_cleanup_all
940 };
941 
942 static inline enum resource_object_create_result
allocate_bo(struct zink_screen * screen,const struct pipe_resource * templ,VkMemoryRequirements * reqs,struct zink_resource_object * obj,struct mem_alloc_info * alloc_info)943 allocate_bo(struct zink_screen *screen, const struct pipe_resource *templ,
944             VkMemoryRequirements *reqs, struct zink_resource_object *obj,
945             struct mem_alloc_info *alloc_info)
946 {
947    VkMemoryAllocateInfo mai;
948    mai.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO;
949    mai.pNext = NULL;
950    mai.allocationSize = reqs->size;
951    enum zink_heap heap = zink_heap_from_domain_flags(alloc_info->flags, alloc_info->aflags);
952    if (templ->flags & PIPE_RESOURCE_FLAG_MAP_COHERENT) {
953       if (!(vk_domain_from_heap(heap) & VK_MEMORY_PROPERTY_HOST_COHERENT_BIT))
954          heap = zink_heap_from_domain_flags(alloc_info->flags & ~VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, alloc_info->aflags);
955    }
956 
957    VkMemoryDedicatedAllocateInfo ded_alloc_info = {
958       .sType = VK_STRUCTURE_TYPE_MEMORY_DEDICATED_ALLOCATE_INFO,
959       .pNext = mai.pNext,
960       .image = obj->image,
961       .buffer = VK_NULL_HANDLE,
962    };
963 
964    if (screen->info.have_KHR_dedicated_allocation && alloc_info->need_dedicated) {
965       ded_alloc_info.pNext = mai.pNext;
966       mai.pNext = &ded_alloc_info;
967    }
968 
969    VkExportMemoryAllocateInfo emai;
970    if ((templ->bind & ZINK_BIND_VIDEO) || ((templ->bind & PIPE_BIND_SHARED) && alloc_info->shared) || (templ->bind & ZINK_BIND_DMABUF)) {
971       emai.sType = VK_STRUCTURE_TYPE_EXPORT_MEMORY_ALLOCATE_INFO;
972       emai.handleTypes = alloc_info->export_types;
973 
974       emai.pNext = mai.pNext;
975       mai.pNext = &emai;
976       obj->exportable = true;
977    }
978 
979 #ifdef ZINK_USE_DMABUF
980 
981 #if !defined(_WIN32)
982    VkImportMemoryFdInfoKHR imfi = {
983       VK_STRUCTURE_TYPE_IMPORT_MEMORY_FD_INFO_KHR,
984       NULL,
985    };
986 
987    if (alloc_info->whandle) {
988       imfi.pNext = NULL;
989       imfi.handleType = alloc_info->external;
990       imfi.fd = os_dupfd_cloexec(alloc_info->whandle->handle);
991       if (imfi.fd < 0) {
992          mesa_loge("ZINK: failed to dup dmabuf fd: %s\n", strerror(errno));
993          return roc_fail_and_cleanup_object;
994       }
995 
996       imfi.pNext = mai.pNext;
997       mai.pNext = &imfi;
998    }
999 #else
1000    VkImportMemoryWin32HandleInfoKHR imfi = {
1001       VK_STRUCTURE_TYPE_IMPORT_MEMORY_WIN32_HANDLE_INFO_KHR,
1002       NULL,
1003    };
1004 
1005    if (alloc_info->whandle) {
1006       HANDLE source_target = GetCurrentProcess();
1007       HANDLE out_handle;
1008 
1009       bool result = DuplicateHandle(source_target, alloc_info->whandle->handle, source_target, &out_handle, 0, false, DUPLICATE_SAME_ACCESS);
1010 
1011       if (!result || !out_handle) {
1012          mesa_loge("ZINK: failed to DuplicateHandle with winerr: %08x\n", (int)GetLastError());
1013          return roc_fail_and_cleanup_object;
1014       }
1015 
1016       imfi.pNext = NULL;
1017       imfi.handleType = alloc_info->external;
1018       imfi.handle = out_handle;
1019 
1020       imfi.pNext = mai.pNext;
1021       mai.pNext = &imfi;
1022    }
1023 #endif
1024 
1025 #endif
1026 
1027    VkImportMemoryHostPointerInfoEXT imhpi = {
1028       VK_STRUCTURE_TYPE_IMPORT_MEMORY_HOST_POINTER_INFO_EXT,
1029       NULL,
1030    };
1031    if (alloc_info->user_mem) {
1032       imhpi.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_HOST_ALLOCATION_BIT_EXT;
1033       imhpi.pHostPointer = (void*)alloc_info->user_mem;
1034       imhpi.pNext = mai.pNext;
1035       mai.pNext = &imhpi;
1036    }
1037 
1038    unsigned alignment = MAX2(reqs->alignment, 256);
1039    if (templ->usage == PIPE_USAGE_STAGING && obj->is_buffer)
1040       alignment = MAX2(alignment, screen->info.props.limits.minMemoryMapAlignment);
1041    obj->alignment = alignment;
1042 
1043    if (zink_mem_type_idx_from_types(screen, heap, reqs->memoryTypeBits) == UINT32_MAX) {
1044       /* not valid based on reqs; demote to more compatible type */
1045       switch (heap) {
1046       case ZINK_HEAP_DEVICE_LOCAL_VISIBLE:
1047          heap = ZINK_HEAP_DEVICE_LOCAL;
1048          break;
1049       case ZINK_HEAP_HOST_VISIBLE_COHERENT_CACHED:
1050          heap = ZINK_HEAP_HOST_VISIBLE_COHERENT;
1051          break;
1052       default:
1053          break;
1054       }
1055       assert(zink_mem_type_idx_from_types(screen, heap, reqs->memoryTypeBits) != UINT32_MAX);
1056    }
1057 
1058    while (1) {
1059       /* iterate over all available memory types to reduce chance of oom */
1060       for (unsigned i = 0; !obj->bo && i < screen->heap_count[heap]; i++) {
1061          if (!(reqs->memoryTypeBits & BITFIELD_BIT(screen->heap_map[heap][i])))
1062             continue;
1063 
1064          mai.memoryTypeIndex = screen->heap_map[heap][i];
1065          obj->bo = zink_bo(zink_bo_create(screen, reqs->size, alignment, heap, mai.pNext ? ZINK_ALLOC_NO_SUBALLOC : 0, mai.memoryTypeIndex, mai.pNext));
1066       }
1067 
1068       if (obj->bo || heap != ZINK_HEAP_DEVICE_LOCAL_VISIBLE)
1069          break;
1070 
1071       /* demote BAR allocations to a different heap on failure to avoid oom */
1072       if (templ->flags & PIPE_RESOURCE_FLAG_MAP_COHERENT || templ->usage == PIPE_USAGE_DYNAMIC)
1073           heap = ZINK_HEAP_HOST_VISIBLE_COHERENT;
1074       else
1075           heap = ZINK_HEAP_DEVICE_LOCAL;
1076    };
1077 
1078    return obj->bo ? roc_success : roc_fail_and_cleanup_object;
1079 }
1080 
1081 static inline bool
update_alloc_info_flags(struct zink_screen * screen,const struct pipe_resource * templ,VkMemoryRequirements * reqs,struct mem_alloc_info * alloc_info)1082 update_alloc_info_flags(struct zink_screen *screen, const struct pipe_resource *templ,
1083                         VkMemoryRequirements *reqs, struct mem_alloc_info *alloc_info)
1084 {
1085    if (templ->flags & PIPE_RESOURCE_FLAG_MAP_COHERENT || templ->usage == PIPE_USAGE_DYNAMIC)
1086       alloc_info->flags |= VK_MEMORY_PROPERTY_HOST_COHERENT_BIT;
1087    else if (!(alloc_info->flags & VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT) &&
1088             templ->usage == PIPE_USAGE_STAGING)
1089       alloc_info->flags |= VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT;
1090 
1091    if (templ->bind & ZINK_BIND_TRANSIENT)
1092       alloc_info->flags |= VK_MEMORY_PROPERTY_LAZILY_ALLOCATED_BIT;
1093 
1094    if (alloc_info->user_mem) {
1095       VkExternalMemoryHandleTypeFlagBits handle_type = VK_EXTERNAL_MEMORY_HANDLE_TYPE_HOST_ALLOCATION_BIT_EXT;
1096       VkMemoryHostPointerPropertiesEXT memory_host_pointer_properties = {0};
1097       memory_host_pointer_properties.sType = VK_STRUCTURE_TYPE_MEMORY_HOST_POINTER_PROPERTIES_EXT;
1098       memory_host_pointer_properties.pNext = NULL;
1099       VkResult res = VKSCR(GetMemoryHostPointerPropertiesEXT)(screen->dev, handle_type, alloc_info->user_mem, &memory_host_pointer_properties);
1100       if (res != VK_SUCCESS) {
1101          mesa_loge("ZINK: vkGetMemoryHostPointerPropertiesEXT failed");
1102          return false;
1103       }
1104       reqs->memoryTypeBits &= memory_host_pointer_properties.memoryTypeBits;
1105       alloc_info->flags &= ~VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
1106    }
1107 
1108    alloc_info->aflags = templ->flags & PIPE_RESOURCE_FLAG_SPARSE ? ZINK_ALLOC_SPARSE : 0;
1109    return true;
1110 }
1111 
1112 static inline void
update_obj_info(struct zink_screen * screen,struct zink_resource_object * obj,const struct pipe_resource * templ,struct mem_alloc_info * alloc_info)1113 update_obj_info(struct zink_screen *screen, struct zink_resource_object *obj,
1114                 const struct pipe_resource *templ, struct mem_alloc_info *alloc_info)
1115 {
1116    if (alloc_info->aflags == ZINK_ALLOC_SPARSE) {
1117       obj->size = templ->width0;
1118    } else {
1119       obj->offset = zink_bo_get_offset(obj->bo);
1120       obj->size = zink_bo_get_size(obj->bo);
1121    }
1122 
1123    obj->coherent = screen->info.mem_props.memoryTypes[obj->bo->base.base.placement].propertyFlags & VK_MEMORY_PROPERTY_HOST_COHERENT_BIT;
1124    if (!(templ->flags & PIPE_RESOURCE_FLAG_SPARSE)) {
1125       obj->host_visible = screen->info.mem_props.memoryTypes[obj->bo->base.base.placement].propertyFlags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
1126    }
1127 }
1128 
1129 static inline void
debug_resource_mem(struct zink_resource_object * obj,const struct pipe_resource * templ,struct zink_screen * screen)1130 debug_resource_mem(struct zink_resource_object *obj, const struct pipe_resource *templ, struct zink_screen *screen)
1131 {
1132       char buf[4096];
1133       unsigned idx = 0;
1134       if (obj->is_buffer) {
1135          size_t size = (size_t)DIV_ROUND_UP(obj->size, 1024);
1136          if (templ->bind == PIPE_BIND_QUERY_BUFFER && templ->usage == PIPE_USAGE_STAGING) //internal qbo
1137             idx += snprintf(buf, sizeof(buf), "QBO(%zu)", size);
1138          else
1139             idx += snprintf(buf, sizeof(buf), "BUF(%zu)", size);
1140       } else {
1141          idx += snprintf(buf, sizeof(buf), "IMG(%s:%ux%ux%u)", util_format_short_name(templ->format), templ->width0, templ->height0, templ->depth0);
1142       }
1143       /*
1144       zink_vkflags_func flag_func = obj->is_buffer ? (zink_vkflags_func)vk_BufferCreateFlagBits_to_str : (zink_vkflags_func)vk_ImageCreateFlagBits_to_str;
1145       zink_vkflags_func usage_func = obj->is_buffer ? (zink_vkflags_func)vk_BufferUsageFlagBits_to_str : (zink_vkflags_func)vk_ImageUsageFlagBits_to_str;
1146       if (obj->vkflags) {
1147          buf[idx++] = '[';
1148          idx += zink_string_vkflags_unroll(&buf[idx], sizeof(buf) - idx, obj->vkflags, flag_func);
1149          buf[idx++] = ']';
1150       }
1151       if (obj->vkusage) {
1152          buf[idx++] = '[';
1153          idx += zink_string_vkflags_unroll(&buf[idx], sizeof(buf) - idx, obj->vkusage, usage_func);
1154          buf[idx++] = ']';
1155       }
1156       */
1157       buf[idx] = 0;
1158       obj->bo->name = zink_debug_mem_add(screen, obj->size, buf);
1159 }
1160 
1161 static inline enum resource_object_create_result
allocate_bo_and_update_obj(struct zink_screen * screen,const struct pipe_resource * templ,VkMemoryRequirements * reqs,struct zink_resource_object * obj,struct mem_alloc_info * alloc_info)1162 allocate_bo_and_update_obj(struct zink_screen *screen, const struct pipe_resource *templ,
1163                            VkMemoryRequirements *reqs, struct zink_resource_object *obj,
1164                            struct mem_alloc_info *alloc_info)
1165 {
1166    if (!update_alloc_info_flags(screen, templ, reqs, alloc_info))
1167       return roc_fail_and_free_object;
1168 
1169    enum resource_object_create_result retval = allocate_bo(screen, templ, reqs, obj, alloc_info);
1170    assert(retval != roc_success_early_return);
1171    if (retval != roc_success)
1172       return retval;
1173 
1174    update_obj_info(screen, obj, templ, alloc_info);
1175 
1176    if (zink_debug & ZINK_DEBUG_MEM)
1177       debug_resource_mem(obj, templ, screen);
1178    return roc_success;
1179 }
1180 
1181 static inline enum resource_object_create_result
create_buffer(struct zink_screen * screen,struct zink_resource_object * obj,const struct pipe_resource * templ,uint64_t * modifiers,int modifiers_count,struct mem_alloc_info * alloc_info)1182 create_buffer(struct zink_screen *screen, struct zink_resource_object *obj,
1183               const struct pipe_resource *templ, uint64_t *modifiers,
1184               int modifiers_count, struct mem_alloc_info *alloc_info)
1185 {
1186    VkBufferCreateInfo bci = create_bci(screen, templ, templ->bind);
1187    VkExternalMemoryBufferCreateInfo embci;
1188    VkMemoryRequirements reqs = {0};
1189 
1190    if (alloc_info->user_mem) {
1191       embci.sType = VK_STRUCTURE_TYPE_EXTERNAL_MEMORY_BUFFER_CREATE_INFO;
1192       embci.pNext = bci.pNext;
1193       embci.handleTypes = VK_EXTERNAL_MEMORY_HANDLE_TYPE_HOST_ALLOCATION_BIT_EXT;
1194       bci.pNext = &embci;
1195    }
1196 
1197    if (VKSCR(CreateBuffer)(screen->dev, &bci, NULL, &obj->buffer) != VK_SUCCESS) {
1198       mesa_loge("ZINK: vkCreateBuffer failed");
1199       return roc_fail_and_free_object;
1200    }
1201 
1202    if (!(templ->bind & (PIPE_BIND_SHADER_IMAGE | ZINK_BIND_DESCRIPTOR))) {
1203        bci.usage |= VK_BUFFER_USAGE_STORAGE_TEXEL_BUFFER_BIT;
1204      if (VKSCR(CreateBuffer)(screen->dev, &bci, NULL, &obj->storage_buffer) != VK_SUCCESS) {
1205         mesa_loge("ZINK: vkCreateBuffer failed");
1206         VKSCR(DestroyBuffer)(screen->dev, obj->buffer, NULL);
1207         return roc_fail_and_free_object;
1208      }
1209    }
1210 
1211    if (modifiers_count) {
1212       assert(modifiers_count == 3);
1213       /* this is the DGC path because there's no other way to pass mem bits and I don't wanna copy/paste everything around */
1214       reqs.size = modifiers[0];
1215       reqs.alignment = modifiers[1];
1216       reqs.memoryTypeBits = modifiers[2];
1217    } else {
1218       VKSCR(GetBufferMemoryRequirements)(screen->dev, obj->buffer, &reqs);
1219    }
1220 
1221    if (templ->usage == PIPE_USAGE_STAGING)
1222       alloc_info->flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT;
1223    else if (templ->usage == PIPE_USAGE_STREAM)
1224       alloc_info->flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
1225    else if (templ->usage == PIPE_USAGE_IMMUTABLE)
1226       alloc_info->flags = VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
1227    else
1228       alloc_info->flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
1229 
1230    obj->is_buffer = true;
1231    obj->transfer_dst = true;
1232    obj->vkflags = bci.flags;
1233    obj->vkusage = bci.usage;
1234 
1235    enum resource_object_create_result retval = allocate_bo_and_update_obj(screen, templ, &reqs, obj,  alloc_info);
1236    assert(retval != roc_success_early_return);
1237    if (retval != roc_success)
1238       return retval;
1239 
1240    if (!(templ->flags & PIPE_RESOURCE_FLAG_SPARSE)) {
1241       if (VKSCR(BindBufferMemory)(screen->dev, obj->buffer, zink_bo_get_mem(obj->bo), obj->offset) != VK_SUCCESS) {
1242          mesa_loge("ZINK: vkBindBufferMemory failed");
1243          return roc_fail_and_cleanup_all ;
1244       }
1245       if (obj->storage_buffer && VKSCR(BindBufferMemory)(screen->dev, obj->storage_buffer, zink_bo_get_mem(obj->bo), obj->offset) != VK_SUCCESS) {
1246          mesa_loge("ZINK: vkBindBufferMemory failed");
1247          return roc_fail_and_cleanup_all;
1248       }
1249    }
1250    return roc_success;
1251 }
1252 
1253 static inline enum resource_object_create_result
create_image(struct zink_screen * screen,struct zink_resource_object * obj,const struct pipe_resource * templ,bool * linear,uint64_t * modifiers,int modifiers_count,struct mem_alloc_info * alloc_info)1254 create_image(struct zink_screen *screen, struct zink_resource_object *obj,
1255              const struct pipe_resource *templ, bool *linear,
1256              uint64_t *modifiers, int modifiers_count,
1257              struct mem_alloc_info *alloc_info)
1258 {
1259    VkMemoryRequirements reqs = {0};
1260    bool winsys_modifier = (alloc_info->export_types & VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT) &&
1261                           alloc_info->whandle &&
1262                           alloc_info->whandle->modifier != DRM_FORMAT_MOD_INVALID;
1263    uint64_t *ici_modifiers = winsys_modifier ? &alloc_info->whandle->modifier : modifiers;
1264    unsigned ici_modifier_count = winsys_modifier ? 1 : modifiers_count;
1265    VkImageCreateInfo ici;
1266    enum pipe_format srgb = PIPE_FORMAT_NONE;
1267    /* we often need to be able to mutate between srgb and linear, but we don't need general
1268     * image view/shader image format compatibility (that path means losing fast clears or compression on some hardware).
1269     */
1270    if (!(templ->bind & ZINK_BIND_MUTABLE)) {
1271       srgb = util_format_is_srgb(templ->format) ? util_format_linear(templ->format) : util_format_srgb(templ->format);
1272       /* why do these helpers have different default return values? */
1273       if (srgb == templ->format)
1274          srgb = PIPE_FORMAT_NONE;
1275    }
1276    VkFormat formats[2];
1277    VkImageFormatListCreateInfo format_list;
1278    if (srgb) {
1279       formats[0] = zink_get_format(screen, templ->format);
1280       formats[1] = zink_get_format(screen, srgb);
1281       /* only use format list if both formats have supported vk equivalents */
1282       if (formats[0] && formats[1]) {
1283          format_list.sType = VK_STRUCTURE_TYPE_IMAGE_FORMAT_LIST_CREATE_INFO;
1284          format_list.pNext = NULL;
1285          format_list.viewFormatCount = 2;
1286          format_list.pViewFormats = formats;
1287          ici.pNext = &format_list;
1288       } else {
1289          ici.pNext = NULL;
1290       }
1291    } else {
1292       ici.pNext = NULL;
1293    }
1294    init_ici(screen, &ici, templ, templ->bind, ici_modifier_count);
1295 
1296    bool success = false;
1297    uint64_t mod = eval_ici(screen, &ici, templ, templ->bind, ici_modifier_count, ici_modifiers, &success);
1298    if (ici.format == VK_FORMAT_A8_UNORM_KHR && !success) {
1299       ici.format = zink_get_format(screen, zink_format_get_emulated_alpha(templ->format));
1300       mod = eval_ici(screen, &ici, templ, templ->bind, ici_modifier_count, ici_modifiers, &success);
1301    }
1302    if (!success)
1303       return roc_fail_and_free_object;
1304 
1305    if (ici.tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT && srgb &&
1306       util_format_get_nr_components(srgb) == 4 &&
1307       !(ici.flags & VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT)) {
1308          mesa_loge("zink: refusing to create possibly-srgb dmabuf due to missing driver support: %s not supported!", util_format_name(srgb));
1309          return roc_fail_and_free_object;
1310    }
1311    VkExternalMemoryImageCreateInfo emici;
1312    VkImageDrmFormatModifierExplicitCreateInfoEXT idfmeci;
1313    VkImageDrmFormatModifierListCreateInfoEXT idfmlci;
1314    VkSubresourceLayout plane_layouts[4];
1315    VkSubresourceLayout plane_layout = {
1316       .offset = alloc_info->whandle ? alloc_info->whandle->offset : 0,
1317       .size = 0,
1318       .rowPitch = alloc_info->whandle ? alloc_info->whandle->stride : 0,
1319       .arrayPitch = 0,
1320       .depthPitch = 0,
1321    };
1322 
1323    obj->render_target = (ici.usage & VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT) != 0;
1324 
1325    if (alloc_info->shared || alloc_info->external) {
1326       emici.sType = VK_STRUCTURE_TYPE_EXTERNAL_MEMORY_IMAGE_CREATE_INFO;
1327       emici.pNext = ici.pNext;
1328       emici.handleTypes = alloc_info->export_types;
1329       ici.pNext = &emici;
1330 
1331       assert(ici.tiling != VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT || mod != DRM_FORMAT_MOD_INVALID);
1332       if (alloc_info->whandle && ici.tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT) {
1333          assert(mod == alloc_info->whandle->modifier || !winsys_modifier);
1334          idfmeci.sType = VK_STRUCTURE_TYPE_IMAGE_DRM_FORMAT_MODIFIER_EXPLICIT_CREATE_INFO_EXT;
1335          idfmeci.pNext = ici.pNext;
1336          idfmeci.drmFormatModifier = mod;
1337          idfmeci.drmFormatModifierPlaneCount = obj->plane_count;
1338 
1339          plane_layouts[0] = plane_layout;
1340          struct pipe_resource *pnext = templ->next;
1341          for (unsigned i = 1; i < obj->plane_count; i++, pnext = pnext->next) {
1342             struct zink_resource *next = zink_resource(pnext);
1343             obj->plane_offsets[i] = plane_layouts[i].offset = next->obj->plane_offsets[i];
1344             obj->plane_strides[i] = plane_layouts[i].rowPitch = next->obj->plane_strides[i];
1345             plane_layouts[i].size = 0;
1346             plane_layouts[i].arrayPitch = 0;
1347             plane_layouts[i].depthPitch = 0;
1348          }
1349          idfmeci.pPlaneLayouts = plane_layouts;
1350 
1351          ici.pNext = &idfmeci;
1352       } else if (ici.tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT) {
1353          idfmlci.sType = VK_STRUCTURE_TYPE_IMAGE_DRM_FORMAT_MODIFIER_LIST_CREATE_INFO_EXT;
1354          idfmlci.pNext = ici.pNext;
1355          idfmlci.drmFormatModifierCount = modifiers_count;
1356          idfmlci.pDrmFormatModifiers = modifiers;
1357          ici.pNext = &idfmlci;
1358       } else if (ici.tiling == VK_IMAGE_TILING_OPTIMAL) {
1359          alloc_info->shared = false;
1360       }
1361    } else if (alloc_info->user_mem) {
1362       emici.sType = VK_STRUCTURE_TYPE_EXTERNAL_MEMORY_IMAGE_CREATE_INFO;
1363       emici.pNext = ici.pNext;
1364       emici.handleTypes = VK_EXTERNAL_MEMORY_HANDLE_TYPE_HOST_ALLOCATION_BIT_EXT;
1365       ici.pNext = &emici;
1366    }
1367 
1368    if (linear)
1369       *linear = ici.tiling == VK_IMAGE_TILING_LINEAR;
1370 
1371    if (ici.usage & VK_IMAGE_USAGE_TRANSFER_DST_BIT)
1372       obj->transfer_dst = true;
1373 
1374 #if defined(ZINK_USE_DMABUF) && !defined(_WIN32)
1375    if (obj->is_aux) {
1376       obj->modifier = mod;
1377       obj->modifier_aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_0_BIT_EXT << alloc_info->whandle->plane;
1378       obj->plane_offsets[alloc_info->whandle->plane] = alloc_info->whandle->offset;
1379       obj->plane_strides[alloc_info->whandle->plane] = alloc_info->whandle->stride;
1380       obj->handle = os_dupfd_cloexec(alloc_info->whandle->handle);
1381       if (obj->handle < 0) {
1382          mesa_loge("ZINK: failed to dup dmabuf fd: %s\n", strerror(errno));
1383          return roc_fail_and_free_object;
1384       }
1385       return roc_success_early_return;
1386    }
1387 #endif
1388 
1389    obj->vkfeats = get_format_feature_flags(ici, screen, templ);;
1390    if (util_format_is_yuv(templ->format)) {
1391       if (!create_sampler_conversion(ici, screen, obj))
1392          return roc_fail_and_free_object;
1393    } else if (alloc_info->whandle) {
1394       obj->plane_strides[alloc_info->whandle->plane] = alloc_info->whandle->stride;
1395    }
1396 
1397    VkResult result = VKSCR(CreateImage)(screen->dev, &ici, NULL, &obj->image);
1398    if (result != VK_SUCCESS) {
1399       mesa_loge("ZINK: vkCreateImage failed (%s)", vk_Result_to_str(result));
1400       return roc_fail_and_free_object;
1401    }
1402 
1403    if (ici.tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT) {
1404       VkImageDrmFormatModifierPropertiesEXT modprops = {0};
1405       modprops.sType = VK_STRUCTURE_TYPE_IMAGE_DRM_FORMAT_MODIFIER_PROPERTIES_EXT;
1406       result = VKSCR(GetImageDrmFormatModifierPropertiesEXT)(screen->dev, obj->image, &modprops);
1407       if (result != VK_SUCCESS) {
1408          mesa_loge("ZINK: vkGetImageDrmFormatModifierPropertiesEXT failed");
1409          return roc_fail_and_free_object;
1410       }
1411       obj->modifier = modprops.drmFormatModifier;
1412       unsigned num_dmabuf_planes = screen->base.get_dmabuf_modifier_planes(&screen->base, obj->modifier, templ->format);
1413       obj->modifier_aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_0_BIT_EXT;
1414       if (num_dmabuf_planes > 1)
1415          obj->modifier_aspect |= VK_IMAGE_ASPECT_MEMORY_PLANE_1_BIT_EXT;
1416       if (num_dmabuf_planes > 2)
1417          obj->modifier_aspect |= VK_IMAGE_ASPECT_MEMORY_PLANE_2_BIT_EXT;
1418       if (num_dmabuf_planes > 3)
1419          obj->modifier_aspect |= VK_IMAGE_ASPECT_MEMORY_PLANE_3_BIT_EXT;
1420       assert(num_dmabuf_planes <= 4);
1421    }
1422 
1423    unsigned num_planes = util_format_get_num_planes(templ->format);
1424    alloc_info->need_dedicated = get_image_memory_requirement(screen, obj, num_planes, &reqs);
1425    if (templ->usage == PIPE_USAGE_STAGING && ici.tiling == VK_IMAGE_TILING_LINEAR)
1426       alloc_info->flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
1427    else
1428       alloc_info->flags = VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
1429 
1430    obj->vkflags = ici.flags;
1431    obj->vkusage = ici.usage;
1432 
1433    enum resource_object_create_result retval = allocate_bo_and_update_obj(screen, templ, &reqs, obj,  alloc_info);
1434    assert(retval != roc_success_early_return);
1435    if (retval != roc_success)
1436       return retval;
1437 
1438    if (num_planes > 1) {
1439       VkBindImageMemoryInfo infos[3];
1440       VkBindImagePlaneMemoryInfo planes[3];
1441       for (unsigned i = 0; i < num_planes; i++) {
1442          infos[i].sType = VK_STRUCTURE_TYPE_BIND_IMAGE_MEMORY_INFO;
1443          infos[i].image = obj->image;
1444          infos[i].memory = zink_bo_get_mem(obj->bo);
1445          infos[i].memoryOffset = obj->plane_offsets[i];
1446          if (templ->bind & ZINK_BIND_VIDEO) {
1447             infos[i].pNext = &planes[i];
1448             planes[i].sType = VK_STRUCTURE_TYPE_BIND_IMAGE_PLANE_MEMORY_INFO;
1449             planes[i].pNext = NULL;
1450             planes[i].planeAspect = plane_aspects[i];
1451          }
1452       }
1453       if (VKSCR(BindImageMemory2)(screen->dev, num_planes, infos) != VK_SUCCESS) {
1454          mesa_loge("ZINK: vkBindImageMemory2 failed");
1455          return roc_fail_and_cleanup_all;
1456       }
1457    } else {
1458       if (!(templ->flags & PIPE_RESOURCE_FLAG_SPARSE))
1459          if (VKSCR(BindImageMemory)(screen->dev, obj->image, zink_bo_get_mem(obj->bo), obj->offset) != VK_SUCCESS) {
1460             mesa_loge("ZINK: vkBindImageMemory failed");
1461             return roc_fail_and_cleanup_all;
1462          }
1463    }
1464 
1465    return roc_success;
1466 }
1467 
1468 static struct zink_resource_object *
resource_object_create(struct zink_screen * screen,const struct pipe_resource * templ,struct winsys_handle * whandle,bool * linear,uint64_t * modifiers,int modifiers_count,const void * loader_private,const void * user_mem)1469 resource_object_create(struct zink_screen *screen, const struct pipe_resource *templ, struct winsys_handle *whandle, bool *linear,
1470                        uint64_t *modifiers, int modifiers_count, const void *loader_private, const void *user_mem)
1471 {
1472    struct zink_resource_object *obj = CALLOC_STRUCT(zink_resource_object);
1473    unsigned max_level = 0;
1474    if (!obj)
1475       return NULL;
1476    simple_mtx_init(&obj->view_lock, mtx_plain);
1477    util_dynarray_init(&obj->views, NULL);
1478    u_rwlock_init(&obj->copy_lock);
1479    obj->unordered_read = true;
1480    obj->unordered_write = true;
1481    obj->unsync_access = true;
1482    obj->last_dt_idx = obj->dt_idx = UINT32_MAX; //TODO: unionize
1483 
1484    struct mem_alloc_info alloc_info = {
1485       .whandle = whandle,
1486       .need_dedicated = false,
1487       .external = 0,
1488       .export_types = ZINK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_BIT,
1489       .shared = templ->bind & PIPE_BIND_SHARED,
1490       .user_mem = user_mem
1491    };
1492 
1493    /* figure out aux plane count */
1494    if (whandle && whandle->plane >= util_format_get_num_planes(whandle->format))
1495       obj->is_aux = true;
1496    struct pipe_resource *pnext = templ->next;
1497    for (obj->plane_count = 1; pnext; obj->plane_count++, pnext = pnext->next) {
1498       struct zink_resource *next = zink_resource(pnext);
1499       if (!next->obj->is_aux)
1500          break;
1501    }
1502 
1503    if (!get_export_flags(screen, templ, &alloc_info)) {
1504       /* can't export anything, fail early */
1505       return NULL;
1506    }
1507 
1508    pipe_reference_init(&obj->reference, 1);
1509    if (loader_private) {
1510       obj->bo = CALLOC_STRUCT(zink_bo);
1511       if (!obj->bo) {
1512          mesa_loge("ZINK: failed to allocate obj->bo!");
1513          return NULL;
1514       }
1515 
1516       obj->transfer_dst = true;
1517       return obj;
1518    }
1519 
1520    enum resource_object_create_result create_result;
1521    if (templ->target == PIPE_BUFFER) {
1522       max_level = 1;
1523       create_result = create_buffer(screen, obj, templ, modifiers, modifiers_count, &alloc_info);
1524    } else {
1525       max_level = templ->last_level + 1;
1526       create_result = create_image(screen, obj, templ, linear, modifiers, modifiers_count,
1527                                    &alloc_info);
1528    }
1529 
1530    switch (create_result) {
1531    case roc_success:
1532       for (unsigned i = 0; i < max_level; i++)
1533          util_dynarray_init(&obj->copies[i], NULL);
1534       FALLTHROUGH;
1535    case roc_success_early_return:
1536       return obj;
1537 
1538    case roc_fail_and_cleanup_all:
1539       zink_bo_unref(screen, obj->bo);
1540       FALLTHROUGH;
1541    case roc_fail_and_cleanup_object:
1542       if (templ->target == PIPE_BUFFER) {
1543          VKSCR(DestroyBuffer)(screen->dev, obj->buffer, NULL);
1544          VKSCR(DestroyBuffer)(screen->dev, obj->storage_buffer, NULL);
1545       } else
1546          VKSCR(DestroyImage)(screen->dev, obj->image, NULL);
1547       FALLTHROUGH;
1548    case roc_fail_and_free_object:
1549       FREE(obj);
1550       return NULL;
1551    default:
1552       unreachable("Invalid create object result code");
1553    }
1554 }
1555 
1556 static struct pipe_resource *
resource_create(struct pipe_screen * pscreen,const struct pipe_resource * templ,struct winsys_handle * whandle,unsigned external_usage,const uint64_t * modifiers,int modifiers_count,const void * loader_private,const void * user_mem)1557 resource_create(struct pipe_screen *pscreen,
1558                 const struct pipe_resource *templ,
1559                 struct winsys_handle *whandle,
1560                 unsigned external_usage,
1561                 const uint64_t *modifiers, int modifiers_count,
1562                 const void *loader_private, const void *user_mem)
1563 {
1564    struct zink_screen *screen = zink_screen(pscreen);
1565    struct zink_resource *res = CALLOC_STRUCT_CL(zink_resource);
1566 
1567    if (!res) {
1568       mesa_loge("ZINK: failed to allocate res!");
1569       return NULL;
1570    }
1571 
1572    if (modifiers_count > 0 && screen->info.have_EXT_image_drm_format_modifier) {
1573       /* for rebinds */
1574       res->modifiers_count = modifiers_count;
1575       res->modifiers = mem_dup(modifiers, modifiers_count * sizeof(uint64_t));
1576       if (!res->modifiers) {
1577          FREE_CL(res);
1578          return NULL;
1579       }
1580    }
1581 
1582    res->base.b = *templ;
1583 
1584    bool allow_cpu_storage = (templ->target == PIPE_BUFFER) &&
1585                             (templ->width0 < 0x1000);
1586    threaded_resource_init(&res->base.b, allow_cpu_storage);
1587    pipe_reference_init(&res->base.b.reference, 1);
1588    res->base.b.screen = pscreen;
1589 
1590    bool linear = false;
1591    struct pipe_resource templ2 = *templ;
1592    if (templ2.flags & PIPE_RESOURCE_FLAG_SPARSE)
1593       templ2.bind |= PIPE_BIND_SHADER_IMAGE;
1594    if (screen->faked_e5sparse && templ->format == PIPE_FORMAT_R9G9B9E5_FLOAT) {
1595       templ2.flags &= ~PIPE_RESOURCE_FLAG_SPARSE;
1596       res->base.b.flags &= ~PIPE_RESOURCE_FLAG_SPARSE;
1597    }
1598    res->obj = resource_object_create(screen, &templ2, whandle, &linear, res->modifiers, res->modifiers_count, loader_private, user_mem);
1599    if (!res->obj) {
1600       free(res->modifiers);
1601       FREE_CL(res);
1602       return NULL;
1603    }
1604 
1605    res->queue = VK_QUEUE_FAMILY_IGNORED;
1606    res->internal_format = templ->format;
1607    if (templ->target == PIPE_BUFFER) {
1608       util_range_init(&res->valid_buffer_range);
1609       res->base.b.bind |= PIPE_BIND_SHADER_IMAGE;
1610       if (!screen->resizable_bar && templ->width0 >= 8196) {
1611          /* We don't want to evict buffers from VRAM by mapping them for CPU access,
1612           * because they might never be moved back again. If a buffer is large enough,
1613           * upload data by copying from a temporary GTT buffer. 8K might not seem much,
1614           * but there can be 100000 buffers.
1615           *
1616           * This tweak improves performance for viewperf.
1617           */
1618          res->base.b.flags |= PIPE_RESOURCE_FLAG_DONT_MAP_DIRECTLY;
1619       }
1620       if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB || zink_debug & ZINK_DEBUG_DGC)
1621          zink_resource_get_address(screen, res);
1622    } else {
1623       if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE)
1624          res->base.b.bind |= PIPE_BIND_SHADER_IMAGE;
1625       if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE) {
1626          uint32_t count = 1;
1627          VKSCR(GetImageSparseMemoryRequirements)(screen->dev, res->obj->image, &count, &res->sparse);
1628          res->base.b.nr_sparse_levels = res->sparse.imageMipTailFirstLod;
1629       }
1630       res->format = zink_get_format(screen, templ->format);
1631       if (templ->target == PIPE_TEXTURE_1D || templ->target == PIPE_TEXTURE_1D_ARRAY) {
1632          res->need_2D = (screen->need_2D_zs && util_format_is_depth_or_stencil(templ->format)) ||
1633                         (screen->need_2D_sparse && (templ->flags & PIPE_RESOURCE_FLAG_SPARSE));
1634       }
1635       res->dmabuf = whandle && whandle->type == WINSYS_HANDLE_TYPE_FD;
1636       if (res->dmabuf)
1637          res->queue = VK_QUEUE_FAMILY_FOREIGN_EXT;
1638       res->layout = res->dmabuf ? VK_IMAGE_LAYOUT_PREINITIALIZED : VK_IMAGE_LAYOUT_UNDEFINED;
1639       res->linear = linear;
1640       res->aspect = aspect_from_format(templ->format);
1641    }
1642 
1643    if (loader_private) {
1644       if (templ->bind & PIPE_BIND_DISPLAY_TARGET) {
1645          /* backbuffer */
1646          res->obj->dt = zink_kopper_displaytarget_create(screen,
1647                                                          res->base.b.bind,
1648                                                          res->base.b.format,
1649                                                          templ->width0,
1650                                                          templ->height0,
1651                                                          64, loader_private,
1652                                                          &res->dt_stride);
1653          if (!res->obj->dt) {
1654             mesa_loge("zink: could not create swapchain");
1655             FREE(res->obj);
1656             free(res->modifiers);
1657             FREE_CL(res);
1658             return NULL;
1659          }
1660          struct kopper_displaytarget *cdt = res->obj->dt;
1661          if (cdt->swapchain->num_acquires) {
1662             /* this should be a reused swapchain after a MakeCurrent dance that deleted the original resource */
1663             for (unsigned i = 0; i < cdt->swapchain->num_images; i++) {
1664                if (!cdt->swapchain->images[i].acquired)
1665                   continue;
1666                res->obj->dt_idx = i;
1667                res->obj->image = cdt->swapchain->images[i].image;
1668                res->layout = cdt->swapchain->images[i].layout;
1669             }
1670          }
1671       } else {
1672          /* frontbuffer */
1673          struct zink_resource *back = (void*)loader_private;
1674          struct kopper_displaytarget *cdt = back->obj->dt;
1675          cdt->refcount++;
1676          assert(back->obj->dt);
1677          res->obj->dt = back->obj->dt;
1678       }
1679       struct kopper_displaytarget *cdt = res->obj->dt;
1680       if (zink_kopper_has_srgb(cdt))
1681          res->obj->vkflags |= VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
1682       if (cdt->swapchain->scci.flags == VK_SWAPCHAIN_CREATE_MUTABLE_FORMAT_BIT_KHR)
1683          res->obj->vkflags = VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT | VK_IMAGE_CREATE_EXTENDED_USAGE_BIT;
1684       res->obj->vkusage = cdt->swapchain->scci.imageUsage;
1685       res->base.b.bind |= PIPE_BIND_DISPLAY_TARGET;
1686       res->linear = false;
1687       res->swapchain = true;
1688    }
1689 
1690    if (!res->obj->host_visible) {
1691       res->base.b.flags |= PIPE_RESOURCE_FLAG_DONT_MAP_DIRECTLY;
1692       res->base.allow_cpu_storage = false;
1693    }
1694    if (res->obj->is_buffer) {
1695       res->base.buffer_id_unique = util_idalloc_mt_alloc(&screen->buffer_ids);
1696       _mesa_hash_table_init(&res->bufferview_cache, NULL, NULL, equals_bvci);
1697       simple_mtx_init(&res->bufferview_mtx, mtx_plain);
1698    } else {
1699       _mesa_hash_table_init(&res->surface_cache, NULL, NULL, equals_ivci);
1700       simple_mtx_init(&res->surface_mtx, mtx_plain);
1701    }
1702    if (res->obj->exportable)
1703       res->base.b.bind |= ZINK_BIND_DMABUF;
1704    return &res->base.b;
1705 }
1706 
1707 static struct pipe_resource *
zink_resource_create(struct pipe_screen * pscreen,const struct pipe_resource * templ)1708 zink_resource_create(struct pipe_screen *pscreen,
1709                      const struct pipe_resource *templ)
1710 {
1711    return resource_create(pscreen, templ, NULL, 0, NULL, 0, NULL, NULL);
1712 }
1713 
1714 static struct pipe_resource *
zink_resource_create_with_modifiers(struct pipe_screen * pscreen,const struct pipe_resource * templ,const uint64_t * modifiers,int modifiers_count)1715 zink_resource_create_with_modifiers(struct pipe_screen *pscreen, const struct pipe_resource *templ,
1716                                     const uint64_t *modifiers, int modifiers_count)
1717 {
1718    return resource_create(pscreen, templ, NULL, 0, modifiers, modifiers_count, NULL, NULL);
1719 }
1720 
1721 static struct pipe_resource *
zink_resource_create_drawable(struct pipe_screen * pscreen,const struct pipe_resource * templ,const void * loader_private)1722 zink_resource_create_drawable(struct pipe_screen *pscreen,
1723                               const struct pipe_resource *templ,
1724                               const void *loader_private)
1725 {
1726    return resource_create(pscreen, templ, NULL, 0, NULL, 0, loader_private, NULL);
1727 }
1728 
1729 static bool
add_resource_bind(struct zink_context * ctx,struct zink_resource * res,unsigned bind)1730 add_resource_bind(struct zink_context *ctx, struct zink_resource *res, unsigned bind)
1731 {
1732    struct zink_screen *screen = zink_screen(ctx->base.screen);
1733    assert((res->base.b.bind & bind) == 0);
1734    res->base.b.bind |= bind;
1735    struct zink_resource_object *old_obj = res->obj;
1736    if (bind & ZINK_BIND_DMABUF && !res->modifiers_count && screen->info.have_EXT_image_drm_format_modifier) {
1737       res->modifiers_count = 1;
1738       res->modifiers = malloc(res->modifiers_count * sizeof(uint64_t));
1739       if (!res->modifiers) {
1740          mesa_loge("ZINK: failed to allocate res->modifiers!");
1741          return false;
1742       }
1743 
1744       res->modifiers[0] = DRM_FORMAT_MOD_LINEAR;
1745    }
1746    struct zink_resource_object *new_obj = resource_object_create(screen, &res->base.b, NULL, &res->linear, res->modifiers, res->modifiers_count, NULL, NULL);
1747    if (!new_obj) {
1748       debug_printf("new backing resource alloc failed!\n");
1749       res->base.b.bind &= ~bind;
1750       return false;
1751    }
1752    struct zink_resource staging = *res;
1753    staging.obj = old_obj;
1754    staging.all_binds = 0;
1755    res->layout = VK_IMAGE_LAYOUT_UNDEFINED;
1756    res->obj = new_obj;
1757    res->queue = VK_QUEUE_FAMILY_IGNORED;
1758    for (unsigned i = 0; i <= res->base.b.last_level; i++) {
1759       struct pipe_box box = {0, 0, 0,
1760                              u_minify(res->base.b.width0, i),
1761                              u_minify(res->base.b.height0, i), res->base.b.array_size};
1762       box.depth = util_num_layers(&res->base.b, i);
1763       ctx->base.resource_copy_region(&ctx->base, &res->base.b, i, 0, 0, 0, &staging.base.b, i, &box);
1764    }
1765    if (old_obj->exportable) {
1766       simple_mtx_lock(&ctx->batch.state->exportable_lock);
1767       _mesa_set_remove_key(&ctx->batch.state->dmabuf_exports, &staging);
1768       simple_mtx_unlock(&ctx->batch.state->exportable_lock);
1769    }
1770    zink_resource_object_reference(screen, &old_obj, NULL);
1771    return true;
1772 }
1773 
1774 static bool
zink_resource_get_param(struct pipe_screen * pscreen,struct pipe_context * pctx,struct pipe_resource * pres,unsigned plane,unsigned layer,unsigned level,enum pipe_resource_param param,unsigned handle_usage,uint64_t * value)1775 zink_resource_get_param(struct pipe_screen *pscreen, struct pipe_context *pctx,
1776                         struct pipe_resource *pres,
1777                         unsigned plane,
1778                         unsigned layer,
1779                         unsigned level,
1780                         enum pipe_resource_param param,
1781                         unsigned handle_usage,
1782                         uint64_t *value)
1783 {
1784    struct zink_screen *screen = zink_screen(pscreen);
1785    struct zink_resource *res = zink_resource(pres);
1786    struct zink_resource_object *obj = res->obj;
1787    struct winsys_handle whandle;
1788    VkImageAspectFlags aspect;
1789    if (obj->modifier_aspect) {
1790       switch (plane) {
1791       case 0:
1792          aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_0_BIT_EXT;
1793          break;
1794       case 1:
1795          aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_1_BIT_EXT;
1796          break;
1797       case 2:
1798          aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_2_BIT_EXT;
1799          break;
1800       case 3:
1801          aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_3_BIT_EXT;
1802          break;
1803       default:
1804          unreachable("how many planes you got in this thing?");
1805       }
1806    } else if (res->obj->sampler_conversion) {
1807       aspect = VK_IMAGE_ASPECT_PLANE_0_BIT;
1808    } else {
1809       aspect = res->aspect;
1810    }
1811    switch (param) {
1812    case PIPE_RESOURCE_PARAM_NPLANES:
1813       if (screen->info.have_EXT_image_drm_format_modifier)
1814          *value = screen->base.get_dmabuf_modifier_planes(&screen->base, obj->modifier, res->internal_format);
1815       else
1816          *value = 1;
1817       break;
1818 
1819    case PIPE_RESOURCE_PARAM_STRIDE: {
1820       VkImageSubresource sub_res = {0};
1821       VkSubresourceLayout sub_res_layout = {0};
1822 
1823       sub_res.aspectMask = aspect;
1824 
1825       VKSCR(GetImageSubresourceLayout)(screen->dev, obj->image, &sub_res, &sub_res_layout);
1826 
1827       *value = sub_res_layout.rowPitch;
1828       break;
1829    }
1830 
1831    case PIPE_RESOURCE_PARAM_OFFSET: {
1832          VkImageSubresource isr = {
1833             aspect,
1834             level,
1835             layer
1836          };
1837          VkSubresourceLayout srl;
1838          VKSCR(GetImageSubresourceLayout)(screen->dev, obj->image, &isr, &srl);
1839          *value = srl.offset;
1840          break;
1841    }
1842 
1843    case PIPE_RESOURCE_PARAM_MODIFIER: {
1844       *value = obj->modifier;
1845       break;
1846    }
1847 
1848    case PIPE_RESOURCE_PARAM_LAYER_STRIDE: {
1849          VkImageSubresource isr = {
1850             aspect,
1851             level,
1852             layer
1853          };
1854          VkSubresourceLayout srl;
1855          VKSCR(GetImageSubresourceLayout)(screen->dev, obj->image, &isr, &srl);
1856          if (res->base.b.target == PIPE_TEXTURE_3D)
1857             *value = srl.depthPitch;
1858          else
1859             *value = srl.arrayPitch;
1860          break;
1861    }
1862 
1863       return false;
1864    case PIPE_RESOURCE_PARAM_HANDLE_TYPE_KMS:
1865    case PIPE_RESOURCE_PARAM_HANDLE_TYPE_SHARED:
1866    case PIPE_RESOURCE_PARAM_HANDLE_TYPE_FD: {
1867 #ifdef ZINK_USE_DMABUF
1868       memset(&whandle, 0, sizeof(whandle));
1869       if (param == PIPE_RESOURCE_PARAM_HANDLE_TYPE_SHARED)
1870          whandle.type = WINSYS_HANDLE_TYPE_SHARED;
1871       if (param == PIPE_RESOURCE_PARAM_HANDLE_TYPE_KMS)
1872          whandle.type = WINSYS_HANDLE_TYPE_KMS;
1873       else if (param == PIPE_RESOURCE_PARAM_HANDLE_TYPE_FD)
1874          whandle.type = WINSYS_HANDLE_TYPE_FD;
1875 
1876       if (!pscreen->resource_get_handle(pscreen, pctx, pres, &whandle, handle_usage))
1877          return false;
1878 
1879 #ifdef _WIN32
1880       *value = (uintptr_t)whandle.handle;
1881 #else
1882       *value = whandle.handle;
1883 #endif
1884       break;
1885 #else
1886       (void)whandle;
1887       return false;
1888 #endif
1889    }
1890    }
1891    return true;
1892 }
1893 
1894 static bool
zink_resource_get_handle(struct pipe_screen * pscreen,struct pipe_context * context,struct pipe_resource * tex,struct winsys_handle * whandle,unsigned usage)1895 zink_resource_get_handle(struct pipe_screen *pscreen,
1896                          struct pipe_context *context,
1897                          struct pipe_resource *tex,
1898                          struct winsys_handle *whandle,
1899                          unsigned usage)
1900 {
1901    if (tex->target == PIPE_BUFFER)
1902       tc_buffer_disable_cpu_storage(tex);
1903    if (whandle->type == WINSYS_HANDLE_TYPE_FD || whandle->type == WINSYS_HANDLE_TYPE_KMS) {
1904 #ifdef ZINK_USE_DMABUF
1905       struct zink_resource *res = zink_resource(tex);
1906       struct zink_screen *screen = zink_screen(pscreen);
1907       struct zink_resource_object *obj = res->obj;
1908 
1909 #if !defined(_WIN32)
1910       if (whandle->type == WINSYS_HANDLE_TYPE_KMS && screen->drm_fd == -1) {
1911          whandle->handle = -1;
1912       } else {
1913          if (!res->obj->exportable) {
1914             assert(!zink_resource_usage_is_unflushed(res));
1915             if (!screen->info.have_EXT_image_drm_format_modifier) {
1916                static bool warned = false;
1917                warn_missing_feature(warned, "EXT_image_drm_format_modifier");
1918                return false;
1919             }
1920             unsigned bind = ZINK_BIND_DMABUF;
1921             if (!(res->base.b.bind & PIPE_BIND_SHARED))
1922                bind |= PIPE_BIND_SHARED;
1923             zink_screen_lock_context(screen);
1924             if (!add_resource_bind(screen->copy_context, res, bind)) {
1925                zink_screen_unlock_context(screen);
1926                return false;
1927             }
1928             if (res->all_binds)
1929                p_atomic_inc(&screen->image_rebind_counter);
1930             screen->copy_context->base.flush(&screen->copy_context->base, NULL, 0);
1931             zink_screen_unlock_context(screen);
1932             obj = res->obj;
1933          }
1934 
1935          VkMemoryGetFdInfoKHR fd_info = {0};
1936          int fd;
1937          fd_info.sType = VK_STRUCTURE_TYPE_MEMORY_GET_FD_INFO_KHR;
1938          fd_info.memory = zink_bo_get_mem(obj->bo);
1939          if (whandle->type == WINSYS_HANDLE_TYPE_FD)
1940             fd_info.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
1941          else
1942             fd_info.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_FD_BIT;
1943          VkResult result = VKSCR(GetMemoryFdKHR)(screen->dev, &fd_info, &fd);
1944          if (result != VK_SUCCESS) {
1945             mesa_loge("ZINK: vkGetMemoryFdKHR failed");
1946             return false;
1947          }
1948          if (whandle->type == WINSYS_HANDLE_TYPE_KMS) {
1949             uint32_t h;
1950             bool ret = zink_bo_get_kms_handle(screen, obj->bo, fd, &h);
1951             close(fd);
1952             if (!ret)
1953                return false;
1954             fd = h;
1955          }
1956 
1957          whandle->handle = fd;
1958       }
1959 #else
1960       VkMemoryGetWin32HandleInfoKHR handle_info = {0};
1961       HANDLE handle;
1962       handle_info.sType = VK_STRUCTURE_TYPE_MEMORY_GET_WIN32_HANDLE_INFO_KHR;
1963       //TODO: remove for wsi
1964       handle_info.memory = zink_bo_get_mem(obj->bo);
1965       handle_info.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_WIN32_BIT;
1966       VkResult result = VKSCR(GetMemoryWin32HandleKHR)(screen->dev, &handle_info, &handle);
1967       if (result != VK_SUCCESS)
1968          return false;
1969       whandle->handle = handle;
1970 #endif
1971       uint64_t value;
1972       zink_resource_get_param(pscreen, context, tex, 0, 0, 0, PIPE_RESOURCE_PARAM_MODIFIER, 0, &value);
1973       whandle->modifier = value;
1974       zink_resource_get_param(pscreen, context, tex, 0, 0, 0, PIPE_RESOURCE_PARAM_OFFSET, 0, &value);
1975       whandle->offset = value;
1976       zink_resource_get_param(pscreen, context, tex, 0, 0, 0, PIPE_RESOURCE_PARAM_STRIDE, 0, &value);
1977       whandle->stride = value;
1978 #else
1979       return false;
1980 #endif
1981    }
1982    return true;
1983 }
1984 
1985 static struct pipe_resource *
zink_resource_from_handle(struct pipe_screen * pscreen,const struct pipe_resource * templ,struct winsys_handle * whandle,unsigned usage)1986 zink_resource_from_handle(struct pipe_screen *pscreen,
1987                  const struct pipe_resource *templ,
1988                  struct winsys_handle *whandle,
1989                  unsigned usage)
1990 {
1991 #ifdef ZINK_USE_DMABUF
1992    if (whandle->modifier != DRM_FORMAT_MOD_INVALID &&
1993        !zink_screen(pscreen)->info.have_EXT_image_drm_format_modifier)
1994       return NULL;
1995 
1996    struct pipe_resource templ2 = *templ;
1997    if (templ->format == PIPE_FORMAT_NONE)
1998       templ2.format = whandle->format;
1999 
2000    uint64_t modifier = DRM_FORMAT_MOD_LINEAR;
2001    int modifier_count = 1;
2002    if (whandle->modifier != DRM_FORMAT_MOD_INVALID)
2003       modifier = whandle->modifier;
2004    else {
2005       if (!zink_screen(pscreen)->driver_workarounds.can_do_invalid_linear_modifier) {
2006          mesa_loge("zink: display server doesn't support DRI3 modifiers and driver can't handle INVALID<->LINEAR!");
2007          return NULL;
2008       }
2009       whandle->modifier = modifier;
2010    }
2011    templ2.bind |= ZINK_BIND_DMABUF;
2012    struct pipe_resource *pres = resource_create(pscreen, &templ2, whandle, usage, &modifier, modifier_count, NULL, NULL);
2013    if (pres) {
2014       struct zink_resource *res = zink_resource(pres);
2015       if (pres->target != PIPE_BUFFER)
2016          res->valid = true;
2017       else
2018          tc_buffer_disable_cpu_storage(pres);
2019       res->internal_format = whandle->format;
2020    }
2021    return pres;
2022 #else
2023    return NULL;
2024 #endif
2025 }
2026 
2027 static struct pipe_resource *
zink_resource_from_user_memory(struct pipe_screen * pscreen,const struct pipe_resource * templ,void * user_memory)2028 zink_resource_from_user_memory(struct pipe_screen *pscreen,
2029                  const struct pipe_resource *templ,
2030                  void *user_memory)
2031 {
2032    struct zink_screen *screen = zink_screen(pscreen);
2033    VkDeviceSize alignMask = screen->info.ext_host_mem_props.minImportedHostPointerAlignment - 1;
2034 
2035    /* Validate the user_memory pointer and fail early.
2036     * minImportedHostPointerAlignment is required to be POT */
2037    if (((uintptr_t)user_memory) & alignMask)
2038       return NULL;
2039 
2040    return resource_create(pscreen, templ, NULL, 0, NULL, 0, NULL, user_memory);
2041 }
2042 
2043 struct zink_memory_object {
2044    struct pipe_memory_object b;
2045    struct winsys_handle whandle;
2046 };
2047 
2048 static struct pipe_memory_object *
zink_memobj_create_from_handle(struct pipe_screen * pscreen,struct winsys_handle * whandle,bool dedicated)2049 zink_memobj_create_from_handle(struct pipe_screen *pscreen, struct winsys_handle *whandle, bool dedicated)
2050 {
2051    struct zink_memory_object *memobj = CALLOC_STRUCT(zink_memory_object);
2052    if (!memobj)
2053       return NULL;
2054    memcpy(&memobj->whandle, whandle, sizeof(struct winsys_handle));
2055    memobj->whandle.type = ZINK_EXTERNAL_MEMORY_HANDLE;
2056 
2057 #ifdef ZINK_USE_DMABUF
2058 
2059 #if !defined(_WIN32)
2060    memobj->whandle.handle = os_dupfd_cloexec(whandle->handle);
2061 #else
2062    HANDLE source_target = GetCurrentProcess();
2063    HANDLE out_handle;
2064 
2065    DuplicateHandle(source_target, whandle->handle, source_target, &out_handle, 0, false, DUPLICATE_SAME_ACCESS);
2066    memobj->whandle.handle = out_handle;
2067 
2068 #endif /* _WIN32 */
2069 #endif /* ZINK_USE_DMABUF */
2070 
2071    return (struct pipe_memory_object *)memobj;
2072 }
2073 
2074 static void
zink_memobj_destroy(struct pipe_screen * pscreen,struct pipe_memory_object * pmemobj)2075 zink_memobj_destroy(struct pipe_screen *pscreen, struct pipe_memory_object *pmemobj)
2076 {
2077 #ifdef ZINK_USE_DMABUF
2078    struct zink_memory_object *memobj = (struct zink_memory_object *)pmemobj;
2079 
2080 #if !defined(_WIN32)
2081    close(memobj->whandle.handle);
2082 #else
2083    CloseHandle(memobj->whandle.handle);
2084 #endif /* _WIN32 */
2085 #endif /* ZINK_USE_DMABUF */
2086 
2087    FREE(pmemobj);
2088 }
2089 
2090 static struct pipe_resource *
zink_resource_from_memobj(struct pipe_screen * pscreen,const struct pipe_resource * templ,struct pipe_memory_object * pmemobj,uint64_t offset)2091 zink_resource_from_memobj(struct pipe_screen *pscreen,
2092                           const struct pipe_resource *templ,
2093                           struct pipe_memory_object *pmemobj,
2094                           uint64_t offset)
2095 {
2096    struct zink_memory_object *memobj = (struct zink_memory_object *)pmemobj;
2097 
2098    struct pipe_resource *pres = resource_create(pscreen, templ, &memobj->whandle, 0, NULL, 0, NULL, NULL);
2099    if (pres) {
2100       if (pres->target != PIPE_BUFFER)
2101          zink_resource(pres)->valid = true;
2102       else
2103          tc_buffer_disable_cpu_storage(pres);
2104    }
2105    return pres;
2106 }
2107 
2108 static bool
invalidate_buffer(struct zink_context * ctx,struct zink_resource * res)2109 invalidate_buffer(struct zink_context *ctx, struct zink_resource *res)
2110 {
2111    struct zink_screen *screen = zink_screen(ctx->base.screen);
2112 
2113    assert(res->base.b.target == PIPE_BUFFER);
2114 
2115    if (res->base.b.flags & PIPE_RESOURCE_FLAG_SPARSE)
2116       return false;
2117 
2118    struct pipe_box box = {0, 0, 0, res->base.b.width0, 0, 0};
2119    if (res->valid_buffer_range.start > res->valid_buffer_range.end &&
2120        !zink_resource_copy_box_intersects(res, 0, &box))
2121       return false;
2122 
2123    if (res->so_valid)
2124       ctx->dirty_so_targets = true;
2125    /* force counter buffer reset */
2126    res->so_valid = false;
2127 
2128    util_range_set_empty(&res->valid_buffer_range);
2129    if (!zink_resource_has_usage(res))
2130       return false;
2131 
2132    struct zink_resource_object *new_obj = resource_object_create(screen, &res->base.b, NULL, NULL, NULL, 0, NULL, 0);
2133    if (!new_obj) {
2134       debug_printf("new backing resource alloc failed!\n");
2135       return false;
2136    }
2137    bool needs_bda = !!res->obj->bda;
2138    /* this ref must be transferred before rebind or else BOOM */
2139    zink_batch_reference_resource_move(&ctx->batch, res);
2140    res->obj = new_obj;
2141    res->queue = VK_QUEUE_FAMILY_IGNORED;
2142    if (needs_bda)
2143       zink_resource_get_address(screen, res);
2144    zink_resource_rebind(ctx, res);
2145    return true;
2146 }
2147 
2148 
2149 static void
zink_resource_invalidate(struct pipe_context * pctx,struct pipe_resource * pres)2150 zink_resource_invalidate(struct pipe_context *pctx, struct pipe_resource *pres)
2151 {
2152    if (pres->target == PIPE_BUFFER)
2153       invalidate_buffer(zink_context(pctx), zink_resource(pres));
2154    else {
2155       struct zink_resource *res = zink_resource(pres);
2156       if (res->valid && res->fb_bind_count)
2157          zink_context(pctx)->rp_loadop_changed = true;
2158       res->valid = false;
2159    }
2160 }
2161 
2162 static void
zink_transfer_copy_bufimage(struct zink_context * ctx,struct zink_resource * dst,struct zink_resource * src,struct zink_transfer * trans)2163 zink_transfer_copy_bufimage(struct zink_context *ctx,
2164                             struct zink_resource *dst,
2165                             struct zink_resource *src,
2166                             struct zink_transfer *trans)
2167 {
2168    assert((trans->base.b.usage & (PIPE_MAP_DEPTH_ONLY | PIPE_MAP_STENCIL_ONLY)) !=
2169           (PIPE_MAP_DEPTH_ONLY | PIPE_MAP_STENCIL_ONLY));
2170 
2171    bool buf2img = src->base.b.target == PIPE_BUFFER;
2172 
2173    struct pipe_box box = trans->base.b.box;
2174    int x = box.x;
2175    if (buf2img)
2176       box.x = trans->offset;
2177 
2178    assert(dst->obj->transfer_dst);
2179    zink_copy_image_buffer(ctx, dst, src, trans->base.b.level, buf2img ? x : 0,
2180                            box.y, box.z, trans->base.b.level, &box, trans->base.b.usage);
2181 }
2182 
2183 ALWAYS_INLINE static void
align_offset_size(const VkDeviceSize alignment,VkDeviceSize * offset,VkDeviceSize * size,VkDeviceSize obj_size)2184 align_offset_size(const VkDeviceSize alignment, VkDeviceSize *offset, VkDeviceSize *size, VkDeviceSize obj_size)
2185 {
2186    VkDeviceSize align = *offset % alignment;
2187    if (alignment - 1 > *offset)
2188       *offset = 0;
2189    else
2190       *offset -= align, *size += align;
2191    align = alignment - (*size % alignment);
2192    if (*offset + *size + align > obj_size)
2193       *size = obj_size - *offset;
2194    else
2195       *size += align;
2196 }
2197 
2198 VkMappedMemoryRange
zink_resource_init_mem_range(struct zink_screen * screen,struct zink_resource_object * obj,VkDeviceSize offset,VkDeviceSize size)2199 zink_resource_init_mem_range(struct zink_screen *screen, struct zink_resource_object *obj, VkDeviceSize offset, VkDeviceSize size)
2200 {
2201    assert(obj->size);
2202    align_offset_size(screen->info.props.limits.nonCoherentAtomSize, &offset, &size, obj->size);
2203    VkMappedMemoryRange range = {
2204       VK_STRUCTURE_TYPE_MAPPED_MEMORY_RANGE,
2205       NULL,
2206       zink_bo_get_mem(obj->bo),
2207       offset,
2208       size
2209    };
2210    assert(range.size);
2211    return range;
2212 }
2213 
2214 static void *
map_resource(struct zink_screen * screen,struct zink_resource * res)2215 map_resource(struct zink_screen *screen, struct zink_resource *res)
2216 {
2217    assert(res->obj->host_visible);
2218    return zink_bo_map(screen, res->obj->bo);
2219 }
2220 
2221 static void
unmap_resource(struct zink_screen * screen,struct zink_resource * res)2222 unmap_resource(struct zink_screen *screen, struct zink_resource *res)
2223 {
2224    zink_bo_unmap(screen, res->obj->bo);
2225 }
2226 
2227 static struct zink_transfer *
create_transfer(struct zink_context * ctx,struct pipe_resource * pres,unsigned usage,const struct pipe_box * box)2228 create_transfer(struct zink_context *ctx, struct pipe_resource *pres, unsigned usage, const struct pipe_box *box)
2229 {
2230    struct zink_transfer *trans;
2231 
2232    if (usage & PIPE_MAP_THREAD_SAFE)
2233       trans = calloc(1, sizeof(*trans));
2234    else if (usage & TC_TRANSFER_MAP_THREADED_UNSYNC)
2235       trans = slab_zalloc(&ctx->transfer_pool_unsync);
2236    else
2237       trans = slab_zalloc(&ctx->transfer_pool);
2238    if (!trans)
2239       return NULL;
2240 
2241    pipe_resource_reference(&trans->base.b.resource, pres);
2242 
2243    trans->base.b.usage = usage;
2244    trans->base.b.box = *box;
2245    return trans;
2246 }
2247 
2248 static void
destroy_transfer(struct zink_context * ctx,struct zink_transfer * trans)2249 destroy_transfer(struct zink_context *ctx, struct zink_transfer *trans)
2250 {
2251    if (trans->base.b.usage & PIPE_MAP_THREAD_SAFE) {
2252       free(trans);
2253    } else {
2254       /* Don't use pool_transfers_unsync. We are always in the driver
2255        * thread. Freeing an object into a different pool is allowed.
2256        */
2257       slab_free(&ctx->transfer_pool, trans);
2258    }
2259 }
2260 
2261 static void *
zink_buffer_map(struct pipe_context * pctx,struct pipe_resource * pres,unsigned level,unsigned usage,const struct pipe_box * box,struct pipe_transfer ** transfer)2262 zink_buffer_map(struct pipe_context *pctx,
2263                     struct pipe_resource *pres,
2264                     unsigned level,
2265                     unsigned usage,
2266                     const struct pipe_box *box,
2267                     struct pipe_transfer **transfer)
2268 {
2269    struct zink_context *ctx = zink_context(pctx);
2270    struct zink_screen *screen = zink_screen(pctx->screen);
2271    struct zink_resource *res = zink_resource(pres);
2272    struct zink_transfer *trans = create_transfer(ctx, pres, usage, box);
2273    if (!trans)
2274       return NULL;
2275 
2276    void *ptr = NULL;
2277 
2278    if (res->base.is_user_ptr)
2279       usage |= PIPE_MAP_PERSISTENT;
2280 
2281    /* See if the buffer range being mapped has never been initialized,
2282     * in which case it can be mapped unsynchronized. */
2283    if (!(usage & (PIPE_MAP_UNSYNCHRONIZED | TC_TRANSFER_MAP_NO_INFER_UNSYNCHRONIZED)) &&
2284        usage & PIPE_MAP_WRITE && !res->base.is_shared &&
2285        !util_ranges_intersect(&res->valid_buffer_range, box->x, box->x + box->width) &&
2286        !zink_resource_copy_box_intersects(res, 0, box)) {
2287       usage |= PIPE_MAP_UNSYNCHRONIZED;
2288    }
2289 
2290    /* If discarding the entire range, discard the whole resource instead. */
2291    if (usage & PIPE_MAP_DISCARD_RANGE && box->x == 0 && box->width == res->base.b.width0) {
2292       usage |= PIPE_MAP_DISCARD_WHOLE_RESOURCE;
2293    }
2294 
2295    /* If a buffer in VRAM is too large and the range is discarded, don't
2296     * map it directly. This makes sure that the buffer stays in VRAM.
2297     */
2298    bool force_discard_range = false;
2299    if (usage & (PIPE_MAP_DISCARD_WHOLE_RESOURCE | PIPE_MAP_DISCARD_RANGE) &&
2300        !(usage & PIPE_MAP_PERSISTENT) &&
2301        res->base.b.flags & PIPE_RESOURCE_FLAG_DONT_MAP_DIRECTLY) {
2302       usage &= ~(PIPE_MAP_DISCARD_WHOLE_RESOURCE | PIPE_MAP_UNSYNCHRONIZED);
2303       usage |= PIPE_MAP_DISCARD_RANGE;
2304       force_discard_range = true;
2305    }
2306 
2307    if (usage & PIPE_MAP_DISCARD_WHOLE_RESOURCE &&
2308        !(usage & (PIPE_MAP_UNSYNCHRONIZED | TC_TRANSFER_MAP_NO_INVALIDATE))) {
2309       assert(usage & PIPE_MAP_WRITE);
2310 
2311       if (invalidate_buffer(ctx, res)) {
2312          /* At this point, the buffer is always idle. */
2313          usage |= PIPE_MAP_UNSYNCHRONIZED;
2314       } else {
2315          /* Fall back to a temporary buffer. */
2316          usage |= PIPE_MAP_DISCARD_RANGE;
2317       }
2318    }
2319 
2320    unsigned map_offset = box->x;
2321    if (usage & PIPE_MAP_DISCARD_RANGE &&
2322         (!res->obj->host_visible ||
2323         !(usage & (PIPE_MAP_UNSYNCHRONIZED | PIPE_MAP_PERSISTENT)))) {
2324 
2325       /* Check if mapping this buffer would cause waiting for the GPU.
2326        */
2327 
2328       if (!res->obj->host_visible || force_discard_range ||
2329           !zink_resource_usage_check_completion(screen, res, ZINK_RESOURCE_ACCESS_RW)) {
2330          /* Do a wait-free write-only transfer using a temporary buffer. */
2331          unsigned offset;
2332 
2333          /* If we are not called from the driver thread, we have
2334           * to use the uploader from u_threaded_context, which is
2335           * local to the calling thread.
2336           */
2337          struct u_upload_mgr *mgr;
2338          if (usage & TC_TRANSFER_MAP_THREADED_UNSYNC)
2339             mgr = ctx->tc->base.stream_uploader;
2340          else
2341             mgr = ctx->base.stream_uploader;
2342          u_upload_alloc(mgr, 0, box->width,
2343                      screen->info.props.limits.minMemoryMapAlignment, &offset,
2344                      (struct pipe_resource **)&trans->staging_res, (void **)&ptr);
2345          res = zink_resource(trans->staging_res);
2346          trans->offset = offset;
2347          usage |= PIPE_MAP_UNSYNCHRONIZED;
2348          ptr = ((uint8_t *)ptr);
2349       } else {
2350          /* At this point, the buffer is always idle (we checked it above). */
2351          usage |= PIPE_MAP_UNSYNCHRONIZED;
2352       }
2353    } else if (usage & PIPE_MAP_DONTBLOCK) {
2354       /* sparse/device-local will always need to wait since it has to copy */
2355       if (!res->obj->host_visible)
2356          goto success;
2357       if (!zink_resource_usage_check_completion(screen, res, ZINK_RESOURCE_ACCESS_WRITE))
2358          goto success;
2359       usage |= PIPE_MAP_UNSYNCHRONIZED;
2360    } else if (!(usage & PIPE_MAP_UNSYNCHRONIZED) &&
2361               (((usage & PIPE_MAP_READ) && !(usage & PIPE_MAP_PERSISTENT) &&
2362                ((screen->info.mem_props.memoryTypes[res->obj->bo->base.base.placement].propertyFlags & VK_STAGING_RAM) != VK_STAGING_RAM)) ||
2363               !res->obj->host_visible)) {
2364       /* the above conditional catches uncached reads and non-HV writes */
2365       assert(!(usage & (TC_TRANSFER_MAP_THREADED_UNSYNC)));
2366       /* any read, non-HV write, or unmappable that reaches this point needs staging */
2367       if ((usage & PIPE_MAP_READ) || !res->obj->host_visible || res->base.b.flags & PIPE_RESOURCE_FLAG_DONT_MAP_DIRECTLY) {
2368 overwrite:
2369          trans->offset = box->x % screen->info.props.limits.minMemoryMapAlignment;
2370          trans->staging_res = pipe_buffer_create(&screen->base, PIPE_BIND_LINEAR, PIPE_USAGE_STAGING, box->width + trans->offset);
2371          if (!trans->staging_res)
2372             goto fail;
2373          struct zink_resource *staging_res = zink_resource(trans->staging_res);
2374          if (usage & PIPE_MAP_THREAD_SAFE) {
2375             /* this map can't access the passed context: use the copy context */
2376             zink_screen_lock_context(screen);
2377             ctx = screen->copy_context;
2378          }
2379          zink_copy_buffer(ctx, staging_res, res, trans->offset, box->x, box->width);
2380          res = staging_res;
2381          usage &= ~PIPE_MAP_UNSYNCHRONIZED;
2382          map_offset = trans->offset;
2383       }
2384    } else if ((usage & PIPE_MAP_UNSYNCHRONIZED) && !res->obj->host_visible) {
2385       trans->offset = box->x % screen->info.props.limits.minMemoryMapAlignment;
2386       trans->staging_res = pipe_buffer_create(&screen->base, PIPE_BIND_LINEAR, PIPE_USAGE_STAGING, box->width + trans->offset);
2387       if (!trans->staging_res)
2388          goto fail;
2389       struct zink_resource *staging_res = zink_resource(trans->staging_res);
2390       res = staging_res;
2391       map_offset = trans->offset;
2392    }
2393 
2394    if (!(usage & PIPE_MAP_UNSYNCHRONIZED)) {
2395       if (usage & PIPE_MAP_WRITE) {
2396          if (!(usage & PIPE_MAP_READ)) {
2397             zink_resource_usage_try_wait(ctx, res, ZINK_RESOURCE_ACCESS_RW);
2398             if (zink_resource_has_unflushed_usage(res))
2399                goto overwrite;
2400          }
2401          zink_resource_usage_wait(ctx, res, ZINK_RESOURCE_ACCESS_RW);
2402       } else
2403          zink_resource_usage_wait(ctx, res, ZINK_RESOURCE_ACCESS_WRITE);
2404       res->obj->access = 0;
2405       res->obj->access_stage = 0;
2406       res->obj->last_write = 0;
2407       zink_resource_copies_reset(res);
2408    }
2409 
2410    if (!ptr) {
2411       /* if writing to a streamout buffer, ensure synchronization next time it's used */
2412       if (usage & PIPE_MAP_WRITE && res->so_valid) {
2413          ctx->dirty_so_targets = true;
2414          /* force counter buffer reset */
2415          res->so_valid = false;
2416       }
2417       ptr = map_resource(screen, res);
2418       if (!ptr)
2419          goto fail;
2420       ptr = ((uint8_t *)ptr) + map_offset;
2421    }
2422 
2423    if (!res->obj->coherent
2424 #if defined(MVK_VERSION)
2425       // Work around for MoltenVk limitation specifically on coherent memory
2426       // MoltenVk returns blank memory ranges when there should be data present
2427       // This is a known limitation of MoltenVK.
2428       // See https://github.com/KhronosGroup/MoltenVK/blob/master/Docs/MoltenVK_Runtime_UserGuide.md#known-moltenvk-limitations
2429 
2430        || screen->instance_info.have_MVK_moltenvk
2431 #endif
2432       ) {
2433       VkDeviceSize size = box->width;
2434       VkDeviceSize offset = res->obj->offset + trans->offset;
2435       VkMappedMemoryRange range = zink_resource_init_mem_range(screen, res->obj, offset, size);
2436       if (VKSCR(InvalidateMappedMemoryRanges)(screen->dev, 1, &range) != VK_SUCCESS) {
2437          mesa_loge("ZINK: vkInvalidateMappedMemoryRanges failed");
2438          zink_bo_unmap(screen, res->obj->bo);
2439          goto fail;
2440       }
2441    }
2442    trans->base.b.usage = usage;
2443    if (usage & PIPE_MAP_WRITE)
2444       util_range_add(&res->base.b, &res->valid_buffer_range, box->x, box->x + box->width);
2445 
2446 success:
2447    /* ensure the copy context gets unlocked */
2448    if (ctx == screen->copy_context)
2449       zink_screen_unlock_context(screen);
2450    *transfer = &trans->base.b;
2451    return ptr;
2452 
2453 fail:
2454    if (ctx == screen->copy_context)
2455       zink_screen_unlock_context(screen);
2456    destroy_transfer(ctx, trans);
2457    return NULL;
2458 }
2459 
2460 static void *
zink_image_map(struct pipe_context * pctx,struct pipe_resource * pres,unsigned level,unsigned usage,const struct pipe_box * box,struct pipe_transfer ** transfer)2461 zink_image_map(struct pipe_context *pctx,
2462                   struct pipe_resource *pres,
2463                   unsigned level,
2464                   unsigned usage,
2465                   const struct pipe_box *box,
2466                   struct pipe_transfer **transfer)
2467 {
2468    struct zink_context *ctx = zink_context(pctx);
2469    struct zink_screen *screen = zink_screen(pctx->screen);
2470    struct zink_resource *res = zink_resource(pres);
2471    struct zink_transfer *trans = create_transfer(ctx, pres, usage, box);
2472    if (!trans)
2473       return NULL;
2474 
2475    trans->base.b.level = level;
2476    if (zink_is_swapchain(res))
2477       /* this is probably a multi-chain which has already been acquired */
2478       zink_kopper_acquire(ctx, res, 0);
2479 
2480    void *ptr;
2481    if (!(usage & PIPE_MAP_UNSYNCHRONIZED)) {
2482       if (usage & PIPE_MAP_WRITE && !(usage & PIPE_MAP_READ))
2483          /* this is like a blit, so we can potentially dump some clears or maybe we have to  */
2484          zink_fb_clears_apply_or_discard(ctx, pres, zink_rect_from_box(box), false);
2485       else if (usage & PIPE_MAP_READ)
2486          /* if the map region intersects with any clears then we have to apply them */
2487          zink_fb_clears_apply_region(ctx, pres, zink_rect_from_box(box));
2488    }
2489    if (!res->linear || !res->obj->host_visible) {
2490       enum pipe_format format = pres->format;
2491       if (usage & PIPE_MAP_DEPTH_ONLY)
2492          format = util_format_get_depth_only(pres->format);
2493       else if (usage & PIPE_MAP_STENCIL_ONLY)
2494          format = PIPE_FORMAT_S8_UINT;
2495       trans->base.b.stride = util_format_get_stride(format, box->width);
2496       trans->base.b.layer_stride = util_format_get_2d_size(format,
2497                                                          trans->base.b.stride,
2498                                                          box->height);
2499 
2500       struct pipe_resource templ = *pres;
2501       templ.next = NULL;
2502       templ.format = format;
2503       templ.usage = usage & PIPE_MAP_READ ? PIPE_USAGE_STAGING : PIPE_USAGE_STREAM;
2504       templ.target = PIPE_BUFFER;
2505       templ.bind = PIPE_BIND_LINEAR;
2506       templ.width0 = trans->base.b.layer_stride * box->depth;
2507       templ.height0 = templ.depth0 = 0;
2508       templ.last_level = 0;
2509       templ.array_size = 1;
2510       templ.flags = 0;
2511 
2512       trans->staging_res = zink_resource_create(pctx->screen, &templ);
2513       if (!trans->staging_res)
2514          goto fail;
2515 
2516       struct zink_resource *staging_res = zink_resource(trans->staging_res);
2517 
2518       if (usage & PIPE_MAP_READ) {
2519          assert(!(usage & TC_TRANSFER_MAP_THREADED_UNSYNC));
2520          /* force multi-context sync */
2521          if (zink_resource_usage_is_unflushed_write(res))
2522             zink_resource_usage_wait(ctx, res, ZINK_RESOURCE_ACCESS_WRITE);
2523          zink_transfer_copy_bufimage(ctx, staging_res, res, trans);
2524          /* need to wait for rendering to finish */
2525          zink_fence_wait(pctx);
2526       }
2527 
2528       ptr = map_resource(screen, staging_res);
2529    } else {
2530       assert(res->linear);
2531       ptr = map_resource(screen, res);
2532       if (!ptr)
2533          goto fail;
2534       if (zink_resource_has_usage(res)) {
2535          assert(!(usage & PIPE_MAP_UNSYNCHRONIZED));
2536          if (usage & PIPE_MAP_WRITE)
2537             zink_fence_wait(pctx);
2538          else
2539             zink_resource_usage_wait(ctx, res, ZINK_RESOURCE_ACCESS_WRITE);
2540       }
2541       VkImageSubresource isr = {
2542          res->modifiers ? res->obj->modifier_aspect : res->aspect,
2543          level,
2544          0
2545       };
2546       VkSubresourceLayout srl;
2547       VKSCR(GetImageSubresourceLayout)(screen->dev, res->obj->image, &isr, &srl);
2548       trans->base.b.stride = srl.rowPitch;
2549       if (res->base.b.target == PIPE_TEXTURE_3D)
2550          trans->base.b.layer_stride = srl.depthPitch;
2551       else
2552          trans->base.b.layer_stride = srl.arrayPitch;
2553       trans->offset = srl.offset;
2554       trans->depthPitch = srl.depthPitch;
2555       const struct util_format_description *desc = util_format_description(res->base.b.format);
2556       unsigned offset = srl.offset +
2557                         box->z * srl.depthPitch +
2558                         (box->y / desc->block.height) * srl.rowPitch +
2559                         (box->x / desc->block.width) * (desc->block.bits / 8);
2560       if (!res->obj->coherent) {
2561          VkDeviceSize size = (VkDeviceSize)box->width * box->height * desc->block.bits / 8;
2562          VkMappedMemoryRange range = zink_resource_init_mem_range(screen, res->obj, res->obj->offset + offset, size);
2563          if (VKSCR(FlushMappedMemoryRanges)(screen->dev, 1, &range) != VK_SUCCESS) {
2564             mesa_loge("ZINK: vkFlushMappedMemoryRanges failed");
2565          }
2566       }
2567       ptr = ((uint8_t *)ptr) + offset;
2568    }
2569    if (!ptr)
2570       goto fail;
2571    if (usage & PIPE_MAP_WRITE) {
2572       if (!res->valid && res->fb_bind_count) {
2573          assert(!(usage & PIPE_MAP_UNSYNCHRONIZED));
2574          ctx->rp_loadop_changed = true;
2575       }
2576       res->valid = true;
2577    }
2578 
2579    if (sizeof(void*) == 4)
2580       trans->base.b.usage |= ZINK_MAP_TEMPORARY;
2581 
2582    *transfer = &trans->base.b;
2583    return ptr;
2584 
2585 fail:
2586    destroy_transfer(ctx, trans);
2587    return NULL;
2588 }
2589 
2590 static void
zink_image_subdata(struct pipe_context * pctx,struct pipe_resource * pres,unsigned level,unsigned usage,const struct pipe_box * box,const void * data,unsigned stride,uintptr_t layer_stride)2591 zink_image_subdata(struct pipe_context *pctx,
2592                   struct pipe_resource *pres,
2593                   unsigned level,
2594                   unsigned usage,
2595                   const struct pipe_box *box,
2596                   const void *data,
2597                   unsigned stride,
2598                   uintptr_t layer_stride)
2599 {
2600    struct zink_screen *screen = zink_screen(pctx->screen);
2601    struct zink_context *ctx = zink_context(pctx);
2602    struct zink_resource *res = zink_resource(pres);
2603 
2604    /* flush clears to avoid subdata conflict */
2605    if (!(usage & TC_TRANSFER_MAP_THREADED_UNSYNC) &&
2606        (res->obj->vkusage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT))
2607       zink_fb_clears_apply_or_discard(ctx, pres, zink_rect_from_box(box), false);
2608    /* only use HIC if supported on image and no pending usage */
2609    while (res->obj->vkusage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT &&
2610           zink_resource_usage_check_completion(screen, res, ZINK_RESOURCE_ACCESS_RW)) {
2611       /* uninit images are always supported */
2612       bool change_layout = res->layout == VK_IMAGE_LAYOUT_UNDEFINED || res->layout == VK_IMAGE_LAYOUT_PREINITIALIZED;
2613       if (!change_layout) {
2614          /* image in some other layout: test for support */
2615          bool can_copy_layout = false;
2616          for (unsigned i = 0; i < screen->info.hic_props.copyDstLayoutCount; i++) {
2617             if (screen->info.hic_props.pCopyDstLayouts[i] == res->layout) {
2618                can_copy_layout = true;
2619                break;
2620             }
2621          }
2622          /* some layouts don't permit HIC copies */
2623          if (!can_copy_layout)
2624             break;
2625       }
2626       bool is_arrayed = false;
2627       switch (pres->target) {
2628       case PIPE_TEXTURE_1D_ARRAY:
2629       case PIPE_TEXTURE_2D_ARRAY:
2630       case PIPE_TEXTURE_CUBE:
2631       case PIPE_TEXTURE_CUBE_ARRAY:
2632          is_arrayed = true;
2633          break;
2634       default: break;
2635       }
2636       /* recalc strides into texel strides because HIC spec is insane */
2637       unsigned vk_stride = util_format_get_stride(pres->format, 1);
2638       stride /= vk_stride;
2639       unsigned vk_layer_stride = util_format_get_2d_size(pres->format, stride, 1) * vk_stride;
2640       layer_stride /= vk_layer_stride;
2641 
2642       VkHostImageLayoutTransitionInfoEXT t = {
2643          VK_STRUCTURE_TYPE_HOST_IMAGE_LAYOUT_TRANSITION_INFO_EXT,
2644          NULL,
2645          res->obj->image,
2646          res->layout,
2647          /* GENERAL support is guaranteed */
2648          VK_IMAGE_LAYOUT_GENERAL,
2649          {res->aspect, 0, VK_REMAINING_MIP_LEVELS, 0, VK_REMAINING_ARRAY_LAYERS}
2650       };
2651       /* only pre-transition uninit images to avoid thrashing */
2652       if (change_layout) {
2653          VKSCR(TransitionImageLayoutEXT)(screen->dev, 1, &t);
2654          res->layout = VK_IMAGE_LAYOUT_GENERAL;
2655       }
2656       VkMemoryToImageCopyEXT region = {
2657          VK_STRUCTURE_TYPE_MEMORY_TO_IMAGE_COPY_EXT,
2658          NULL,
2659          data,
2660          stride,
2661          layer_stride,
2662          {res->aspect, level, is_arrayed ? box->z : 0, is_arrayed ? box->depth : 1},
2663          {box->x, box->y, is_arrayed ? 0 : box->z},
2664          {box->width, box->height, is_arrayed ? 1 : box->depth}
2665       };
2666       VkCopyMemoryToImageInfoEXT copy = {
2667          VK_STRUCTURE_TYPE_COPY_MEMORY_TO_IMAGE_INFO_EXT,
2668          NULL,
2669          0,
2670          res->obj->image,
2671          res->layout,
2672          1,
2673          &region
2674       };
2675       VKSCR(CopyMemoryToImageEXT)(screen->dev, &copy);
2676       if (change_layout && screen->can_hic_shader_read && !pres->last_level && !box->x && !box->y && !box->z &&
2677           box->width == pres->width0 && box->height == pres->height0 &&
2678           ((is_arrayed && box->depth == pres->array_size) || (!is_arrayed && box->depth == pres->depth0))) {
2679          /* assume full copy single-mip images use shader read access */
2680          t.oldLayout = VK_IMAGE_LAYOUT_GENERAL;
2681          t.newLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
2682          VKSCR(TransitionImageLayoutEXT)(screen->dev, 1, &t);
2683          res->layout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
2684          /* assume multi-mip where further subdata calls may happen */
2685       }
2686       /* make sure image is marked as having data */
2687       res->valid = true;
2688       return;
2689    }
2690    /* fallback case for per-resource unsupported or device-level unsupported */
2691    u_default_texture_subdata(pctx, pres, level, usage, box, data, stride, layer_stride);
2692 }
2693 
2694 static void
zink_transfer_flush_region(struct pipe_context * pctx,struct pipe_transfer * ptrans,const struct pipe_box * box)2695 zink_transfer_flush_region(struct pipe_context *pctx,
2696                            struct pipe_transfer *ptrans,
2697                            const struct pipe_box *box)
2698 {
2699    struct zink_context *ctx = zink_context(pctx);
2700    struct zink_resource *res = zink_resource(ptrans->resource);
2701    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
2702 
2703    if (trans->base.b.usage & PIPE_MAP_WRITE) {
2704       struct zink_screen *screen = zink_screen(pctx->screen);
2705       struct zink_resource *m = trans->staging_res ? zink_resource(trans->staging_res) :
2706                                                      res;
2707       ASSERTED VkDeviceSize size, src_offset, dst_offset = 0;
2708       if (m->obj->is_buffer) {
2709          size = box->width;
2710          src_offset = box->x + (trans->staging_res ? trans->offset : ptrans->box.x);
2711          dst_offset = box->x + ptrans->box.x;
2712       } else {
2713          size = (VkDeviceSize)box->width * box->height * util_format_get_blocksize(m->base.b.format);
2714          src_offset = trans->offset +
2715                   box->z * trans->depthPitch +
2716                   util_format_get_2d_size(m->base.b.format, trans->base.b.stride, box->y) +
2717                   util_format_get_stride(m->base.b.format, box->x);
2718          assert(src_offset + size <= res->obj->size);
2719       }
2720       if (!m->obj->coherent) {
2721          VkMappedMemoryRange range = zink_resource_init_mem_range(screen, m->obj, m->obj->offset, m->obj->size);
2722          if (VKSCR(FlushMappedMemoryRanges)(screen->dev, 1, &range) != VK_SUCCESS) {
2723             mesa_loge("ZINK: vkFlushMappedMemoryRanges failed");
2724          }
2725       }
2726       if (trans->staging_res) {
2727          struct zink_resource *staging_res = zink_resource(trans->staging_res);
2728 
2729          if (ptrans->resource->target == PIPE_BUFFER)
2730             zink_copy_buffer(ctx, res, staging_res, dst_offset, src_offset, size);
2731          else
2732             zink_transfer_copy_bufimage(ctx, res, staging_res, trans);
2733       }
2734    }
2735 }
2736 
2737 /* used to determine whether to emit a TRANSFER_DST barrier on copies */
2738 bool
zink_resource_copy_box_intersects(struct zink_resource * res,unsigned level,const struct pipe_box * box)2739 zink_resource_copy_box_intersects(struct zink_resource *res, unsigned level, const struct pipe_box *box)
2740 {
2741    /* if there are no valid copy rects tracked, this needs a barrier */
2742    if (!res->obj->copies_valid)
2743       return true;
2744    /* untracked huge miplevel */
2745    if (level >= ARRAY_SIZE(res->obj->copies))
2746       return true;
2747    u_rwlock_rdlock(&res->obj->copy_lock);
2748    struct pipe_box *b = res->obj->copies[level].data;
2749    unsigned num_boxes = util_dynarray_num_elements(&res->obj->copies[level], struct pipe_box);
2750    bool (*intersect)(const struct pipe_box *, const struct pipe_box *);
2751    /* determine intersection function based on dimensionality */
2752    switch (res->base.b.target) {
2753    case PIPE_BUFFER:
2754    case PIPE_TEXTURE_1D:
2755       intersect = u_box_test_intersection_1d;
2756       break;
2757 
2758    case PIPE_TEXTURE_1D_ARRAY:
2759    case PIPE_TEXTURE_2D:
2760       intersect = u_box_test_intersection_2d;
2761       break;
2762 
2763    default:
2764       intersect = u_box_test_intersection_3d;
2765       break;
2766    }
2767    /* if any of the tracked boxes intersect with this one, a barrier is needed */
2768    bool ret = false;
2769    for (unsigned i = 0; i < num_boxes; i++) {
2770       if (intersect(box, b + i)) {
2771          ret = true;
2772          break;
2773       }
2774    }
2775    u_rwlock_rdunlock(&res->obj->copy_lock);
2776    /* no intersection = no barrier */
2777    return ret;
2778 }
2779 
2780 /* track a new region for TRANSFER_DST barrier emission */
2781 void
zink_resource_copy_box_add(struct zink_context * ctx,struct zink_resource * res,unsigned level,const struct pipe_box * box)2782 zink_resource_copy_box_add(struct zink_context *ctx, struct zink_resource *res, unsigned level, const struct pipe_box *box)
2783 {
2784    u_rwlock_wrlock(&res->obj->copy_lock);
2785    if (res->obj->copies_valid) {
2786       struct pipe_box *b = res->obj->copies[level].data;
2787       unsigned num_boxes = util_dynarray_num_elements(&res->obj->copies[level], struct pipe_box);
2788       for (unsigned i = 0; i < num_boxes; i++) {
2789          switch (res->base.b.target) {
2790          case PIPE_BUFFER:
2791          case PIPE_TEXTURE_1D:
2792             /* no-op included region */
2793             if (b[i].x <= box->x && b[i].x + b[i].width >= box->x + box->width)
2794                goto out;
2795 
2796             /* try to merge adjacent regions */
2797             if (b[i].x == box->x + box->width) {
2798                b[i].x -= box->width;
2799                b[i].width += box->width;
2800                goto out;
2801             }
2802             if (b[i].x + b[i].width == box->x) {
2803                b[i].width += box->width;
2804                goto out;
2805             }
2806 
2807             /* try to merge into region */
2808             if (box->x <= b[i].x && box->x + box->width >= b[i].x + b[i].width) {
2809                *b = *box;
2810                goto out;
2811             }
2812             break;
2813 
2814          case PIPE_TEXTURE_1D_ARRAY:
2815          case PIPE_TEXTURE_2D:
2816             /* no-op included region */
2817             if (b[i].x <= box->x && b[i].x + b[i].width >= box->x + box->width &&
2818                 b[i].y <= box->y && b[i].y + b[i].height >= box->y + box->height)
2819                goto out;
2820 
2821             /* try to merge adjacent regions */
2822             if (b[i].y == box->y && b[i].height == box->height) {
2823                if (b[i].x == box->x + box->width) {
2824                   b[i].x -= box->width;
2825                   b[i].width += box->width;
2826                   goto out;
2827                }
2828                if (b[i].x + b[i].width == box->x) {
2829                   b[i].width += box->width;
2830                   goto out;
2831                }
2832             } else if (b[i].x == box->x && b[i].width == box->width) {
2833                if (b[i].y == box->y + box->height) {
2834                   b[i].y -= box->height;
2835                   b[i].height += box->height;
2836                   goto out;
2837                }
2838                if (b[i].y + b[i].height == box->y) {
2839                   b[i].height += box->height;
2840                   goto out;
2841                }
2842             }
2843 
2844             /* try to merge into region */
2845             if (box->x <= b[i].x && box->x + box->width >= b[i].x + b[i].width &&
2846                 box->y <= b[i].y && box->y + box->height >= b[i].y + b[i].height) {
2847                *b = *box;
2848                goto out;
2849             }
2850             break;
2851 
2852          default:
2853             /* no-op included region */
2854             if (b[i].x <= box->x && b[i].x + b[i].width >= box->x + box->width &&
2855                 b[i].y <= box->y && b[i].y + b[i].height >= box->y + box->height &&
2856                 b[i].z <= box->z && b[i].z + b[i].depth >= box->z + box->depth)
2857                goto out;
2858 
2859                /* try to merge adjacent regions */
2860             if (b[i].z == box->z && b[i].depth == box->depth) {
2861                if (b[i].y == box->y && b[i].height == box->height) {
2862                   if (b[i].x == box->x + box->width) {
2863                      b[i].x -= box->width;
2864                      b[i].width += box->width;
2865                      goto out;
2866                   }
2867                   if (b[i].x + b[i].width == box->x) {
2868                      b[i].width += box->width;
2869                      goto out;
2870                   }
2871                } else if (b[i].x == box->x && b[i].width == box->width) {
2872                   if (b[i].y == box->y + box->height) {
2873                      b[i].y -= box->height;
2874                      b[i].height += box->height;
2875                      goto out;
2876                   }
2877                   if (b[i].y + b[i].height == box->y) {
2878                      b[i].height += box->height;
2879                      goto out;
2880                   }
2881                }
2882             } else if (b[i].x == box->x && b[i].width == box->width) {
2883                if (b[i].y == box->y && b[i].height == box->height) {
2884                   if (b[i].z == box->z + box->depth) {
2885                      b[i].z -= box->depth;
2886                      b[i].depth += box->depth;
2887                      goto out;
2888                   }
2889                   if (b[i].z + b[i].depth == box->z) {
2890                      b[i].depth += box->depth;
2891                      goto out;
2892                   }
2893                } else if (b[i].z == box->z && b[i].depth == box->depth) {
2894                   if (b[i].y == box->y + box->height) {
2895                      b[i].y -= box->height;
2896                      b[i].height += box->height;
2897                      goto out;
2898                   }
2899                   if (b[i].y + b[i].height == box->y) {
2900                      b[i].height += box->height;
2901                      goto out;
2902                   }
2903                }
2904             } else if (b[i].y == box->y && b[i].height == box->height) {
2905                if (b[i].z == box->z && b[i].depth == box->depth) {
2906                   if (b[i].x == box->x + box->width) {
2907                      b[i].x -= box->width;
2908                      b[i].width += box->width;
2909                      goto out;
2910                   }
2911                   if (b[i].x + b[i].width == box->x) {
2912                      b[i].width += box->width;
2913                      goto out;
2914                   }
2915                } else if (b[i].x == box->x && b[i].width == box->width) {
2916                   if (b[i].z == box->z + box->depth) {
2917                      b[i].z -= box->depth;
2918                      b[i].depth += box->depth;
2919                      goto out;
2920                   }
2921                   if (b[i].z + b[i].depth == box->z) {
2922                      b[i].depth += box->depth;
2923                      goto out;
2924                   }
2925                }
2926             }
2927 
2928             /* try to merge into region */
2929             if (box->x <= b[i].x && box->x + box->width >= b[i].x + b[i].width &&
2930                 box->y <= b[i].y && box->y + box->height >= b[i].y + b[i].height &&
2931                 box->z <= b[i].z && box->z + box->depth >= b[i].z + b[i].depth)
2932                goto out;
2933 
2934             break;
2935          }
2936       }
2937    }
2938    util_dynarray_append(&res->obj->copies[level], struct pipe_box, *box);
2939    if (!res->copies_warned && util_dynarray_num_elements(&res->obj->copies[level], struct pipe_box) > 100) {
2940       perf_debug(ctx, "zink: PERF WARNING! > 100 copy boxes detected for %p\n", res);
2941       mesa_logw("zink: PERF WARNING! > 100 copy boxes detected for %p\n", res);
2942       res->copies_warned = true;
2943    }
2944    res->obj->copies_valid = true;
2945 out:
2946    u_rwlock_wrunlock(&res->obj->copy_lock);
2947 }
2948 
2949 void
zink_resource_copies_reset(struct zink_resource * res)2950 zink_resource_copies_reset(struct zink_resource *res)
2951 {
2952    if (!res->obj->copies_valid)
2953       return;
2954    u_rwlock_wrlock(&res->obj->copy_lock);
2955    unsigned max_level = res->base.b.target == PIPE_BUFFER ? 1 : (res->base.b.last_level + 1);
2956    if (res->base.b.target == PIPE_BUFFER) {
2957       /* flush transfer regions back to valid range on reset */
2958       struct pipe_box *b = res->obj->copies[0].data;
2959       unsigned num_boxes = util_dynarray_num_elements(&res->obj->copies[0], struct pipe_box);
2960       for (unsigned i = 0; i < num_boxes; i++)
2961          util_range_add(&res->base.b, &res->valid_buffer_range, b[i].x, b[i].x + b[i].width);
2962    }
2963    for (unsigned i = 0; i < max_level; i++)
2964       util_dynarray_clear(&res->obj->copies[i]);
2965    res->obj->copies_valid = false;
2966    res->obj->copies_need_reset = false;
2967    u_rwlock_wrunlock(&res->obj->copy_lock);
2968 }
2969 
2970 static void
transfer_unmap(struct pipe_context * pctx,struct pipe_transfer * ptrans)2971 transfer_unmap(struct pipe_context *pctx, struct pipe_transfer *ptrans)
2972 {
2973    struct zink_context *ctx = zink_context(pctx);
2974    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
2975 
2976    if (!(trans->base.b.usage & (PIPE_MAP_FLUSH_EXPLICIT | PIPE_MAP_COHERENT))) {
2977       /* flush_region is relative to the mapped region: use only the extents */
2978       struct pipe_box box = ptrans->box;
2979       box.x = box.y = box.z = 0;
2980       zink_transfer_flush_region(pctx, ptrans, &box);
2981    }
2982 
2983    if (trans->staging_res)
2984       pipe_resource_reference(&trans->staging_res, NULL);
2985    pipe_resource_reference(&trans->base.b.resource, NULL);
2986 
2987    destroy_transfer(ctx, trans);
2988 }
2989 
2990 static void
do_transfer_unmap(struct zink_screen * screen,struct zink_transfer * trans)2991 do_transfer_unmap(struct zink_screen *screen, struct zink_transfer *trans)
2992 {
2993    struct zink_resource *res = zink_resource(trans->staging_res);
2994    if (!res)
2995       res = zink_resource(trans->base.b.resource);
2996    unmap_resource(screen, res);
2997 }
2998 
2999 void
zink_screen_buffer_unmap(struct pipe_screen * pscreen,struct pipe_transfer * ptrans)3000 zink_screen_buffer_unmap(struct pipe_screen *pscreen, struct pipe_transfer *ptrans)
3001 {
3002    struct zink_screen *screen = zink_screen(pscreen);
3003    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
3004    if (trans->base.b.usage & PIPE_MAP_ONCE && !trans->staging_res)
3005       do_transfer_unmap(screen, trans);
3006    transfer_unmap(NULL, ptrans);
3007 }
3008 
3009 static void
zink_buffer_unmap(struct pipe_context * pctx,struct pipe_transfer * ptrans)3010 zink_buffer_unmap(struct pipe_context *pctx, struct pipe_transfer *ptrans)
3011 {
3012    struct zink_screen *screen = zink_screen(pctx->screen);
3013    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
3014    if (trans->base.b.usage & PIPE_MAP_ONCE && !trans->staging_res)
3015       do_transfer_unmap(screen, trans);
3016    transfer_unmap(pctx, ptrans);
3017 }
3018 
3019 static void
zink_image_unmap(struct pipe_context * pctx,struct pipe_transfer * ptrans)3020 zink_image_unmap(struct pipe_context *pctx, struct pipe_transfer *ptrans)
3021 {
3022    struct zink_screen *screen = zink_screen(pctx->screen);
3023    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
3024    if (sizeof(void*) == 4)
3025       do_transfer_unmap(screen, trans);
3026    transfer_unmap(pctx, ptrans);
3027 }
3028 
3029 static void
zink_buffer_subdata(struct pipe_context * ctx,struct pipe_resource * buffer,unsigned usage,unsigned offset,unsigned size,const void * data)3030 zink_buffer_subdata(struct pipe_context *ctx, struct pipe_resource *buffer,
3031                     unsigned usage, unsigned offset, unsigned size, const void *data)
3032 {
3033    struct pipe_transfer *transfer = NULL;
3034    struct pipe_box box;
3035    uint8_t *map = NULL;
3036 
3037    usage |= PIPE_MAP_WRITE;
3038 
3039    if (!(usage & PIPE_MAP_DIRECTLY))
3040       usage |= PIPE_MAP_DISCARD_RANGE;
3041 
3042    u_box_1d(offset, size, &box);
3043    map = zink_buffer_map(ctx, buffer, 0, usage, &box, &transfer);
3044    if (!map)
3045       return;
3046 
3047    memcpy(map, data, size);
3048    zink_buffer_unmap(ctx, transfer);
3049 }
3050 
3051 static struct pipe_resource *
zink_resource_get_separate_stencil(struct pipe_resource * pres)3052 zink_resource_get_separate_stencil(struct pipe_resource *pres)
3053 {
3054    /* For packed depth-stencil, we treat depth as the primary resource
3055     * and store S8 as the "second plane" resource.
3056     */
3057    if (pres->next && pres->next->format == PIPE_FORMAT_S8_UINT)
3058       return pres->next;
3059 
3060    return NULL;
3061 
3062 }
3063 
3064 static bool
resource_object_add_bind(struct zink_context * ctx,struct zink_resource * res,unsigned bind)3065 resource_object_add_bind(struct zink_context *ctx, struct zink_resource *res, unsigned bind)
3066 {
3067    /* base resource already has the cap */
3068    if (res->base.b.bind & bind)
3069       return true;
3070    if (res->obj->is_buffer) {
3071       unreachable("zink: all buffers should have this bit");
3072       return true;
3073    }
3074    assert(!res->obj->dt);
3075    zink_fb_clears_apply_region(ctx, &res->base.b, (struct u_rect){0, res->base.b.width0, 0, res->base.b.height0});
3076    bool ret = add_resource_bind(ctx, res, bind);
3077    if (ret)
3078       zink_resource_rebind(ctx, res);
3079 
3080    return ret;
3081 }
3082 
3083 bool
zink_resource_object_init_storage(struct zink_context * ctx,struct zink_resource * res)3084 zink_resource_object_init_storage(struct zink_context *ctx, struct zink_resource *res)
3085 {
3086    return resource_object_add_bind(ctx, res, PIPE_BIND_SHADER_IMAGE);
3087 }
3088 
3089 bool
zink_resource_object_init_mutable(struct zink_context * ctx,struct zink_resource * res)3090 zink_resource_object_init_mutable(struct zink_context *ctx, struct zink_resource *res)
3091 {
3092    return resource_object_add_bind(ctx, res, ZINK_BIND_MUTABLE);
3093 }
3094 
3095 VkDeviceAddress
zink_resource_get_address(struct zink_screen * screen,struct zink_resource * res)3096 zink_resource_get_address(struct zink_screen *screen, struct zink_resource *res)
3097 {
3098    assert(res->obj->is_buffer);
3099    if (!res->obj->bda) {
3100       VkBufferDeviceAddressInfo info = {
3101          VK_STRUCTURE_TYPE_BUFFER_DEVICE_ADDRESS_INFO,
3102          NULL,
3103          res->obj->buffer
3104       };
3105       res->obj->bda = VKSCR(GetBufferDeviceAddress)(screen->dev, &info);
3106    }
3107    return res->obj->bda;
3108 }
3109 
3110 void
zink_resource_setup_transfer_layouts(struct zink_context * ctx,struct zink_resource * src,struct zink_resource * dst)3111 zink_resource_setup_transfer_layouts(struct zink_context *ctx, struct zink_resource *src, struct zink_resource *dst)
3112 {
3113    if (src == dst) {
3114       /* The Vulkan 1.1 specification says the following about valid usage
3115        * of vkCmdBlitImage:
3116        *
3117        * "srcImageLayout must be VK_IMAGE_LAYOUT_SHARED_PRESENT_KHR,
3118        *  VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL or VK_IMAGE_LAYOUT_GENERAL"
3119        *
3120        * and:
3121        *
3122        * "dstImageLayout must be VK_IMAGE_LAYOUT_SHARED_PRESENT_KHR,
3123        *  VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL or VK_IMAGE_LAYOUT_GENERAL"
3124        *
3125        * Since we cant have the same image in two states at the same time,
3126        * we're effectively left with VK_IMAGE_LAYOUT_SHARED_PRESENT_KHR or
3127        * VK_IMAGE_LAYOUT_GENERAL. And since this isn't a present-related
3128        * operation, VK_IMAGE_LAYOUT_GENERAL seems most appropriate.
3129        */
3130       zink_screen(ctx->base.screen)->image_barrier(ctx, src,
3131                                   VK_IMAGE_LAYOUT_GENERAL,
3132                                   VK_ACCESS_TRANSFER_READ_BIT | VK_ACCESS_TRANSFER_WRITE_BIT,
3133                                   VK_PIPELINE_STAGE_TRANSFER_BIT);
3134    } else {
3135       zink_screen(ctx->base.screen)->image_barrier(ctx, src,
3136                                   VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
3137                                   VK_ACCESS_TRANSFER_READ_BIT,
3138                                   VK_PIPELINE_STAGE_TRANSFER_BIT);
3139 
3140       zink_screen(ctx->base.screen)->image_barrier(ctx, dst,
3141                                   VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL,
3142                                   VK_ACCESS_TRANSFER_WRITE_BIT,
3143                                   VK_PIPELINE_STAGE_TRANSFER_BIT);
3144    }
3145 }
3146 
3147 void
zink_get_depth_stencil_resources(struct pipe_resource * res,struct zink_resource ** out_z,struct zink_resource ** out_s)3148 zink_get_depth_stencil_resources(struct pipe_resource *res,
3149                                  struct zink_resource **out_z,
3150                                  struct zink_resource **out_s)
3151 {
3152    if (!res) {
3153       if (out_z) *out_z = NULL;
3154       if (out_s) *out_s = NULL;
3155       return;
3156    }
3157 
3158    if (res->format != PIPE_FORMAT_S8_UINT) {
3159       if (out_z) *out_z = zink_resource(res);
3160       if (out_s) *out_s = zink_resource(zink_resource_get_separate_stencil(res));
3161    } else {
3162       if (out_z) *out_z = NULL;
3163       if (out_s) *out_s = zink_resource(res);
3164    }
3165 }
3166 
3167 static void
zink_resource_set_separate_stencil(struct pipe_resource * pres,struct pipe_resource * stencil)3168 zink_resource_set_separate_stencil(struct pipe_resource *pres,
3169                                    struct pipe_resource *stencil)
3170 {
3171    assert(util_format_has_depth(util_format_description(pres->format)));
3172    pipe_resource_reference(&pres->next, stencil);
3173 }
3174 
3175 static enum pipe_format
zink_resource_get_internal_format(struct pipe_resource * pres)3176 zink_resource_get_internal_format(struct pipe_resource *pres)
3177 {
3178    struct zink_resource *res = zink_resource(pres);
3179    return res->internal_format;
3180 }
3181 
3182 static const struct u_transfer_vtbl transfer_vtbl = {
3183    .resource_create       = zink_resource_create,
3184    .resource_destroy      = zink_resource_destroy,
3185    .transfer_map          = zink_image_map,
3186    .transfer_unmap        = zink_image_unmap,
3187    .transfer_flush_region = zink_transfer_flush_region,
3188    .get_internal_format   = zink_resource_get_internal_format,
3189    .set_stencil           = zink_resource_set_separate_stencil,
3190    .get_stencil           = zink_resource_get_separate_stencil,
3191 };
3192 
3193 bool
zink_screen_resource_init(struct pipe_screen * pscreen)3194 zink_screen_resource_init(struct pipe_screen *pscreen)
3195 {
3196    struct zink_screen *screen = zink_screen(pscreen);
3197    pscreen->resource_create = u_transfer_helper_resource_create;
3198    pscreen->resource_create_with_modifiers = zink_resource_create_with_modifiers;
3199    pscreen->resource_create_drawable = zink_resource_create_drawable;
3200    pscreen->resource_destroy = u_transfer_helper_resource_destroy;
3201    pscreen->transfer_helper = u_transfer_helper_create(&transfer_vtbl,
3202       U_TRANSFER_HELPER_SEPARATE_Z32S8 | U_TRANSFER_HELPER_SEPARATE_STENCIL |
3203       U_TRANSFER_HELPER_INTERLEAVE_IN_PLACE |
3204       U_TRANSFER_HELPER_MSAA_MAP |
3205       (!screen->have_D24_UNORM_S8_UINT ? U_TRANSFER_HELPER_Z24_IN_Z32F : 0));
3206 
3207    if (screen->info.have_KHR_external_memory_fd || screen->info.have_KHR_external_memory_win32) {
3208       pscreen->resource_get_handle = zink_resource_get_handle;
3209       pscreen->resource_from_handle = zink_resource_from_handle;
3210    }
3211    if (screen->info.have_EXT_external_memory_host) {
3212       pscreen->resource_from_user_memory = zink_resource_from_user_memory;
3213    }
3214    if (screen->instance_info.have_KHR_external_memory_capabilities) {
3215       pscreen->memobj_create_from_handle = zink_memobj_create_from_handle;
3216       pscreen->memobj_destroy = zink_memobj_destroy;
3217       pscreen->resource_from_memobj = zink_resource_from_memobj;
3218    }
3219    pscreen->resource_get_param = zink_resource_get_param;
3220    return true;
3221 }
3222 
3223 void
zink_context_resource_init(struct pipe_context * pctx)3224 zink_context_resource_init(struct pipe_context *pctx)
3225 {
3226    pctx->buffer_map = zink_buffer_map;
3227    pctx->buffer_unmap = zink_buffer_unmap;
3228    pctx->texture_map = u_transfer_helper_transfer_map;
3229    pctx->texture_unmap = u_transfer_helper_transfer_unmap;
3230 
3231    pctx->transfer_flush_region = u_transfer_helper_transfer_flush_region;
3232    pctx->buffer_subdata = zink_buffer_subdata;
3233    pctx->texture_subdata = zink_image_subdata;
3234    pctx->invalidate_resource = zink_resource_invalidate;
3235 }
3236