• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /*
2  * Copyright 2018 Collabora Ltd.
3  *
4  * Permission is hereby granted, free of charge, to any person obtaining a
5  * copy of this software and associated documentation files (the "Software"),
6  * to deal in the Software without restriction, including without limitation
7  * on the rights to use, copy, modify, merge, publish, distribute, sub
8  * license, and/or sell copies of the Software, and to permit persons to whom
9  * the Software is furnished to do so, subject to the following conditions:
10  *
11  * The above copyright notice and this permission notice (including the next
12  * paragraph) shall be included in all copies or substantial portions of the
13  * Software.
14  *
15  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17  * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
18  * THE AUTHOR(S) AND/OR THEIR SUPPLIERS BE LIABLE FOR ANY CLAIM,
19  * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
20  * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
21  * USE OR OTHER DEALINGS IN THE SOFTWARE.
22  */
23 
24 #include "zink_resource.h"
25 
26 #include "zink_batch.h"
27 #include "zink_clear.h"
28 #include "zink_context.h"
29 #include "zink_fence.h"
30 #include "zink_format.h"
31 #include "zink_program.h"
32 #include "zink_screen.h"
33 #include "zink_kopper.h"
34 
35 #ifdef VK_USE_PLATFORM_METAL_EXT
36 #include "QuartzCore/CAMetalLayer.h"
37 #endif
38 
39 #include "vk_format.h"
40 #include "util/u_blitter.h"
41 #include "util/u_debug.h"
42 #include "util/format/u_format.h"
43 #include "util/u_transfer_helper.h"
44 #include "util/u_inlines.h"
45 #include "util/u_memory.h"
46 #include "util/u_upload_mgr.h"
47 #include "util/os_file.h"
48 #include "frontend/winsys_handle.h"
49 
50 #if !defined(__APPLE__)
51 #define ZINK_USE_DMABUF
52 #endif
53 
54 #if defined(ZINK_USE_DMABUF) && !defined(_WIN32)
55 #include "drm-uapi/drm_fourcc.h"
56 #else
57 /* these won't actually be used */
58 #define DRM_FORMAT_MOD_INVALID 0
59 #define DRM_FORMAT_MOD_LINEAR 0
60 #endif
61 
62 #ifdef __APPLE__
63 #include "MoltenVK/mvk_vulkan.h"
64 // Source of MVK_VERSION
65 #include "MoltenVK/mvk_config.h"
66 #endif /* __APPLE__ */
67 
68 #define ZINK_EXTERNAL_MEMORY_HANDLE 999
69 
70 
71 
72 struct zink_debug_mem_entry {
73    uint32_t count;
74    uint64_t size;
75    const char *name;
76 };
77 
78 static const char *
zink_debug_mem_add(struct zink_screen * screen,uint64_t size,const char * name)79 zink_debug_mem_add(struct zink_screen *screen, uint64_t size, const char *name)
80 {
81    assert(name);
82 
83    simple_mtx_lock(&screen->debug_mem_lock);
84    struct hash_entry *entry = _mesa_hash_table_search(screen->debug_mem_sizes, name);
85    struct zink_debug_mem_entry *debug_bos;
86 
87    if (!entry) {
88       debug_bos = calloc(1, sizeof(struct zink_debug_mem_entry));
89       debug_bos->name = strdup(name);
90       _mesa_hash_table_insert(screen->debug_mem_sizes, debug_bos->name, debug_bos);
91    } else {
92       debug_bos = (struct zink_debug_mem_entry *) entry->data;
93    }
94 
95    debug_bos->count++;
96    debug_bos->size += align(size, 4096);
97    simple_mtx_unlock(&screen->debug_mem_lock);
98 
99    return debug_bos->name;
100 }
101 
102 static void
zink_debug_mem_del(struct zink_screen * screen,struct zink_bo * bo)103 zink_debug_mem_del(struct zink_screen *screen, struct zink_bo *bo)
104 {
105    simple_mtx_lock(&screen->debug_mem_lock);
106    struct hash_entry *entry = _mesa_hash_table_search(screen->debug_mem_sizes, bo->name);
107    /* If we're finishing the BO, it should have been added already */
108    assert(entry);
109 
110    struct zink_debug_mem_entry *debug_bos = entry->data;
111    debug_bos->count--;
112    debug_bos->size -= align(zink_bo_get_size(bo), 4096);
113    if (!debug_bos->count) {
114       _mesa_hash_table_remove(screen->debug_mem_sizes, entry);
115       free((void*)debug_bos->name);
116       free(debug_bos);
117    }
118    simple_mtx_unlock(&screen->debug_mem_lock);
119 }
120 
121 static int
debug_bos_count_compare(const void * in_a,const void * in_b)122 debug_bos_count_compare(const void *in_a, const void *in_b)
123 {
124    struct zink_debug_mem_entry *a = *(struct zink_debug_mem_entry **)in_a;
125    struct zink_debug_mem_entry *b = *(struct zink_debug_mem_entry **)in_b;
126    return a->count - b->count;
127 }
128 
129 void
zink_debug_mem_print_stats(struct zink_screen * screen)130 zink_debug_mem_print_stats(struct zink_screen *screen)
131 {
132    simple_mtx_lock(&screen->debug_mem_lock);
133 
134    /* Put the HT's sizes data in an array so we can sort by number of allocations. */
135    struct util_dynarray dyn;
136    util_dynarray_init(&dyn, NULL);
137 
138    uint32_t size = 0;
139    uint32_t count = 0;
140    hash_table_foreach(screen->debug_mem_sizes, entry)
141    {
142       struct zink_debug_mem_entry *debug_bos = entry->data;
143       util_dynarray_append(&dyn, struct zink_debug_mem_entry *, debug_bos);
144       size += debug_bos->size / 1024;
145       count += debug_bos->count;
146    }
147 
148    qsort(dyn.data,
149          util_dynarray_num_elements(&dyn, struct zink_debug_mem_entry *),
150          sizeof(struct zink_debug_mem_entryos_entry *), debug_bos_count_compare);
151 
152    util_dynarray_foreach(&dyn, struct zink_debug_mem_entry *, entryp)
153    {
154       struct zink_debug_mem_entry *debug_bos = *entryp;
155       mesa_logi("%30s: %4d bos, %lld kb\n", debug_bos->name, debug_bos->count,
156                 (long long) (debug_bos->size / 1024));
157    }
158 
159    mesa_logi("submitted %d bos (%d MB)\n", count, DIV_ROUND_UP(size, 1024));
160 
161    util_dynarray_fini(&dyn);
162 
163    simple_mtx_unlock(&screen->debug_mem_lock);
164 }
165 
166 static bool
equals_ivci(const void * a,const void * b)167 equals_ivci(const void *a, const void *b)
168 {
169    const uint8_t *pa = a;
170    const uint8_t *pb = b;
171    size_t offset = offsetof(VkImageViewCreateInfo, flags);
172    return memcmp(pa + offset, pb + offset, sizeof(VkImageViewCreateInfo) - offset) == 0;
173 }
174 
175 static bool
equals_bvci(const void * a,const void * b)176 equals_bvci(const void *a, const void *b)
177 {
178    const uint8_t *pa = a;
179    const uint8_t *pb = b;
180    size_t offset = offsetof(VkBufferViewCreateInfo, flags);
181    return memcmp(pa + offset, pb + offset, sizeof(VkBufferViewCreateInfo) - offset) == 0;
182 }
183 
184 static void
185 zink_transfer_flush_region(struct pipe_context *pctx,
186                            struct pipe_transfer *ptrans,
187                            const struct pipe_box *box);
188 
189 void
debug_describe_zink_resource_object(char * buf,const struct zink_resource_object * ptr)190 debug_describe_zink_resource_object(char *buf, const struct zink_resource_object *ptr)
191 {
192    sprintf(buf, "zink_resource_object");
193 }
194 
195 void
zink_destroy_resource_object(struct zink_screen * screen,struct zink_resource_object * obj)196 zink_destroy_resource_object(struct zink_screen *screen, struct zink_resource_object *obj)
197 {
198    if (obj->is_buffer) {
199       while (util_dynarray_contains(&obj->views, VkBufferView))
200          VKSCR(DestroyBufferView)(screen->dev, util_dynarray_pop(&obj->views, VkBufferView), NULL);
201    } else {
202       while (util_dynarray_contains(&obj->views, VkImageView))
203          VKSCR(DestroyImageView)(screen->dev, util_dynarray_pop(&obj->views, VkImageView), NULL);
204    }
205    if (!obj->dt && zink_debug & ZINK_DEBUG_MEM)
206       zink_debug_mem_del(screen, obj->bo);
207    util_dynarray_fini(&obj->views);
208    for (unsigned i = 0; i < ARRAY_SIZE(obj->copies); i++)
209       util_dynarray_fini(&obj->copies[i]);
210    if (obj->is_buffer) {
211       VKSCR(DestroyBuffer)(screen->dev, obj->buffer, NULL);
212       VKSCR(DestroyBuffer)(screen->dev, obj->storage_buffer, NULL);
213    } else if (obj->dt) {
214       zink_kopper_displaytarget_destroy(screen, obj->dt);
215    } else if (!obj->is_aux) {
216       VKSCR(DestroyImage)(screen->dev, obj->image, NULL);
217    } else {
218 #if defined(ZINK_USE_DMABUF) && !defined(_WIN32)
219       close(obj->handle);
220 #endif
221    }
222 
223    simple_mtx_destroy(&obj->view_lock);
224    if (obj->dt) {
225       FREE(obj->bo); //this is a dummy struct
226    } else
227       zink_bo_unref(screen, obj->bo);
228    FREE(obj);
229 }
230 
231 static void
zink_resource_destroy(struct pipe_screen * pscreen,struct pipe_resource * pres)232 zink_resource_destroy(struct pipe_screen *pscreen,
233                       struct pipe_resource *pres)
234 {
235    struct zink_screen *screen = zink_screen(pscreen);
236    struct zink_resource *res = zink_resource(pres);
237    /* prevent double-free when unrefing internal surfaces */
238    res->base.b.reference.count = 999;
239    if (pres->target == PIPE_BUFFER) {
240       util_range_destroy(&res->valid_buffer_range);
241       util_idalloc_mt_free(&screen->buffer_ids, res->base.buffer_id_unique);
242       assert(!_mesa_hash_table_num_entries(&res->bufferview_cache));
243       simple_mtx_destroy(&res->bufferview_mtx);
244       ralloc_free(res->bufferview_cache.table);
245    } else {
246       pipe_surface_reference(&res->surface, NULL);
247       assert(!_mesa_hash_table_num_entries(&res->surface_cache));
248       simple_mtx_destroy(&res->surface_mtx);
249       ralloc_free(res->surface_cache.table);
250    }
251    /* no need to do anything for the caches, these objects own the resource lifetimes */
252 
253    free(res->modifiers);
254    zink_resource_object_reference(screen, &res->obj, NULL);
255    threaded_resource_deinit(pres);
256    FREE_CL(res);
257 }
258 
259 static VkImageAspectFlags
aspect_from_format(enum pipe_format fmt)260 aspect_from_format(enum pipe_format fmt)
261 {
262    if (util_format_is_depth_or_stencil(fmt)) {
263       VkImageAspectFlags aspect = 0;
264       const struct util_format_description *desc = util_format_description(fmt);
265       if (util_format_has_depth(desc))
266          aspect |= VK_IMAGE_ASPECT_DEPTH_BIT;
267       if (util_format_has_stencil(desc))
268          aspect |= VK_IMAGE_ASPECT_STENCIL_BIT;
269       return aspect;
270    } else
271      return VK_IMAGE_ASPECT_COLOR_BIT;
272 }
273 
274 static VkBufferCreateInfo
create_bci(struct zink_screen * screen,const struct pipe_resource * templ,unsigned bind)275 create_bci(struct zink_screen *screen, const struct pipe_resource *templ, unsigned bind)
276 {
277    VkBufferCreateInfo bci;
278    bci.sType = VK_STRUCTURE_TYPE_BUFFER_CREATE_INFO;
279    bci.pNext = NULL;
280    bci.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
281    bci.queueFamilyIndexCount = 0;
282    bci.pQueueFamilyIndices = NULL;
283    bci.size = templ->width0;
284    bci.flags = 0;
285    assert(bci.size > 0);
286 
287    if (bind & ZINK_BIND_DESCRIPTOR) {
288       /* gallium sizes are all uint32_t, while the total size of this buffer may exceed that limit */
289       bci.usage = 0;
290       bci.usage |= VK_BUFFER_USAGE_SAMPLER_DESCRIPTOR_BUFFER_BIT_EXT |
291                    VK_BUFFER_USAGE_RESOURCE_DESCRIPTOR_BUFFER_BIT_EXT;
292    } else {
293       bci.usage = VK_BUFFER_USAGE_TRANSFER_SRC_BIT |
294                   VK_BUFFER_USAGE_TRANSFER_DST_BIT |
295                   VK_BUFFER_USAGE_STORAGE_BUFFER_BIT;
296 
297       bci.usage |= VK_BUFFER_USAGE_UNIFORM_TEXEL_BUFFER_BIT |
298                   VK_BUFFER_USAGE_INDIRECT_BUFFER_BIT |
299                   VK_BUFFER_USAGE_VERTEX_BUFFER_BIT |
300                   VK_BUFFER_USAGE_INDEX_BUFFER_BIT |
301                   VK_BUFFER_USAGE_UNIFORM_BUFFER_BIT;
302 
303       if (screen->info.have_EXT_transform_feedback) {
304          bci.usage |= VK_BUFFER_USAGE_TRANSFORM_FEEDBACK_BUFFER_BIT_EXT |
305                       VK_BUFFER_USAGE_TRANSFORM_FEEDBACK_COUNTER_BUFFER_BIT_EXT;
306       }
307    }
308    if (screen->info.have_KHR_buffer_device_address)
309       bci.usage |= VK_BUFFER_USAGE_SHADER_DEVICE_ADDRESS_BIT;
310 
311    if (bind & PIPE_BIND_SHADER_IMAGE)
312       bci.usage |= VK_BUFFER_USAGE_STORAGE_TEXEL_BUFFER_BIT;
313 
314    if (bind & PIPE_BIND_QUERY_BUFFER)
315       bci.usage |= VK_BUFFER_USAGE_CONDITIONAL_RENDERING_BIT_EXT;
316 
317    if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE)
318       bci.flags |= VK_BUFFER_CREATE_SPARSE_BINDING_BIT | VK_BUFFER_CREATE_SPARSE_RESIDENCY_BIT;
319    return bci;
320 }
321 
322 static bool
check_ici(struct zink_screen * screen,VkImageCreateInfo * ici,uint64_t modifier)323 check_ici(struct zink_screen *screen, VkImageCreateInfo *ici, uint64_t modifier)
324 {
325    VkImageFormatProperties image_props;
326    VkResult ret;
327    bool optimalDeviceAccess = true;
328    assert(modifier == DRM_FORMAT_MOD_INVALID ||
329           (VKSCR(GetPhysicalDeviceImageFormatProperties2) && screen->info.have_EXT_image_drm_format_modifier));
330    if (VKSCR(GetPhysicalDeviceImageFormatProperties2)) {
331       VkImageFormatProperties2 props2;
332       props2.sType = VK_STRUCTURE_TYPE_IMAGE_FORMAT_PROPERTIES_2;
333       props2.pNext = NULL;
334       VkSamplerYcbcrConversionImageFormatProperties ycbcr_props;
335       ycbcr_props.sType = VK_STRUCTURE_TYPE_SAMPLER_YCBCR_CONVERSION_IMAGE_FORMAT_PROPERTIES;
336       ycbcr_props.pNext = NULL;
337       if (screen->info.have_KHR_sampler_ycbcr_conversion)
338          props2.pNext = &ycbcr_props;
339       VkHostImageCopyDevicePerformanceQueryEXT hic = {
340          VK_STRUCTURE_TYPE_HOST_IMAGE_COPY_DEVICE_PERFORMANCE_QUERY_EXT,
341          props2.pNext,
342       };
343       if (screen->info.have_EXT_host_image_copy && ici->usage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT)
344          props2.pNext = &hic;
345       VkPhysicalDeviceImageFormatInfo2 info;
346       info.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_IMAGE_FORMAT_INFO_2;
347       /* possibly VkImageFormatListCreateInfo */
348       info.pNext = ici->pNext;
349       info.format = ici->format;
350       info.type = ici->imageType;
351       info.tiling = ici->tiling;
352       info.usage = ici->usage;
353       info.flags = ici->flags;
354 
355       VkPhysicalDeviceImageDrmFormatModifierInfoEXT mod_info;
356       if (modifier != DRM_FORMAT_MOD_INVALID) {
357          mod_info.sType = VK_STRUCTURE_TYPE_PHYSICAL_DEVICE_IMAGE_DRM_FORMAT_MODIFIER_INFO_EXT;
358          mod_info.pNext = info.pNext;
359          mod_info.drmFormatModifier = modifier;
360          mod_info.sharingMode = VK_SHARING_MODE_EXCLUSIVE;
361          mod_info.queueFamilyIndexCount = 0;
362          mod_info.pQueueFamilyIndices = NULL;
363          info.pNext = &mod_info;
364       }
365 
366       ret = VKSCR(GetPhysicalDeviceImageFormatProperties2)(screen->pdev, &info, &props2);
367       image_props = props2.imageFormatProperties;
368       if (screen->info.have_EXT_host_image_copy && ici->usage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT)
369          optimalDeviceAccess = hic.optimalDeviceAccess;
370 #if defined(MVK_VERSION)
371       // MoltenVK cannot allocate a depth buffer with VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT
372       // As hic.identicalMemoryLayout is set, it is not needed?
373       if (screen->info.have_EXT_host_image_copy && ici->usage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT && hic.identicalMemoryLayout)
374          optimalDeviceAccess = false;
375 #endif
376    } else
377       ret = VKSCR(GetPhysicalDeviceImageFormatProperties)(screen->pdev, ici->format, ici->imageType,
378                                                    ici->tiling, ici->usage, ici->flags, &image_props);
379    if (ret != VK_SUCCESS)
380       return false;
381    if (ici->extent.depth > image_props.maxExtent.depth ||
382        ici->extent.height > image_props.maxExtent.height ||
383        ici->extent.width > image_props.maxExtent.width)
384       return false;
385    if (ici->mipLevels > image_props.maxMipLevels)
386       return false;
387    if (ici->arrayLayers > image_props.maxArrayLayers)
388       return false;
389    if (!(ici->samples & image_props.sampleCounts))
390       return false;
391    if (!optimalDeviceAccess)
392       return false;
393    return true;
394 }
395 
396 static VkImageUsageFlags
get_image_usage_for_feats(struct zink_screen * screen,VkFormatFeatureFlags2 feats,const struct pipe_resource * templ,unsigned bind,bool * need_extended)397 get_image_usage_for_feats(struct zink_screen *screen, VkFormatFeatureFlags2 feats, const struct pipe_resource *templ, unsigned bind, bool *need_extended)
398 {
399    VkImageUsageFlags usage = 0;
400    bool is_planar = util_format_get_num_planes(templ->format) > 1;
401    *need_extended = false;
402 
403    if (bind & ZINK_BIND_TRANSIENT)
404       usage |= VK_IMAGE_USAGE_TRANSIENT_ATTACHMENT_BIT;
405    else {
406       /* sadly, gallium doesn't let us know if it'll ever need this, so we have to assume */
407       if (is_planar || (feats & VK_FORMAT_FEATURE_TRANSFER_SRC_BIT))
408          usage |= VK_IMAGE_USAGE_TRANSFER_SRC_BIT;
409       if (is_planar || (feats & VK_FORMAT_FEATURE_TRANSFER_DST_BIT))
410          usage |= VK_IMAGE_USAGE_TRANSFER_DST_BIT;
411       if (feats & VK_FORMAT_FEATURE_SAMPLED_IMAGE_BIT)
412          usage |= VK_IMAGE_USAGE_SAMPLED_BIT;
413 
414       if ((is_planar || (feats & VK_FORMAT_FEATURE_STORAGE_IMAGE_BIT)) && (bind & PIPE_BIND_SHADER_IMAGE)) {
415          assert(templ->nr_samples <= 1 || screen->info.feats.features.shaderStorageImageMultisample);
416          usage |= VK_IMAGE_USAGE_STORAGE_BIT;
417       }
418    }
419 
420    if (bind & PIPE_BIND_RENDER_TARGET) {
421       if (feats & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT) {
422          usage |= VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
423          if (!(bind & ZINK_BIND_TRANSIENT) && (bind & (PIPE_BIND_LINEAR | PIPE_BIND_SHARED)) != (PIPE_BIND_LINEAR | PIPE_BIND_SHARED))
424             usage |= VK_IMAGE_USAGE_INPUT_ATTACHMENT_BIT;
425          if (!(bind & ZINK_BIND_TRANSIENT) && screen->info.have_EXT_attachment_feedback_loop_layout)
426             usage |= VK_IMAGE_USAGE_ATTACHMENT_FEEDBACK_LOOP_BIT_EXT;
427       } else {
428          /* trust that gallium isn't going to give us anything wild */
429          *need_extended = true;
430          return 0;
431       }
432    } else if ((bind & PIPE_BIND_SAMPLER_VIEW) && !util_format_is_depth_or_stencil(templ->format)) {
433       if (!(feats & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT)) {
434          /* ensure we can u_blitter this later */
435          *need_extended = true;
436          return 0;
437       }
438       usage |= VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
439    }
440 
441    if (bind & PIPE_BIND_DEPTH_STENCIL) {
442       if (feats & VK_FORMAT_FEATURE_DEPTH_STENCIL_ATTACHMENT_BIT)
443          usage |= VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT;
444       else
445          return 0;
446       if (screen->info.have_EXT_attachment_feedback_loop_layout && !(bind & ZINK_BIND_TRANSIENT))
447          usage |= VK_IMAGE_USAGE_ATTACHMENT_FEEDBACK_LOOP_BIT_EXT;
448    /* this is unlikely to occur and has been included for completeness */
449    } else if (bind & PIPE_BIND_SAMPLER_VIEW && !(usage & VK_IMAGE_USAGE_TRANSFER_DST_BIT)) {
450       if (feats & VK_FORMAT_FEATURE_COLOR_ATTACHMENT_BIT)
451          usage |= VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
452       else
453          return 0;
454    }
455 
456    if (bind & PIPE_BIND_STREAM_OUTPUT)
457       usage |= VK_IMAGE_USAGE_INPUT_ATTACHMENT_BIT;
458 
459    /* Add host transfer if not sparse */
460    if (!(templ->flags & PIPE_RESOURCE_FLAG_SPARSE) &&
461        screen->info.have_EXT_host_image_copy &&
462        feats & VK_FORMAT_FEATURE_2_HOST_IMAGE_TRANSFER_BIT_EXT)
463       usage |= VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT;
464 
465    return usage;
466 }
467 
468 static VkFormatFeatureFlags
find_modifier_feats(const struct zink_modifier_props * prop,uint64_t modifier)469 find_modifier_feats(const struct zink_modifier_props *prop, uint64_t modifier)
470 {
471    for (unsigned j = 0; j < prop->drmFormatModifierCount; j++) {
472       if (prop->pDrmFormatModifierProperties[j].drmFormatModifier == modifier) {
473          return prop->pDrmFormatModifierProperties[j].drmFormatModifierTilingFeatures;
474       }
475    }
476    return 0;
477 }
478 
479 /* check HIC optimalness */
480 static bool
suboptimal_check_ici(struct zink_screen * screen,VkImageCreateInfo * ici,uint64_t mod)481 suboptimal_check_ici(struct zink_screen *screen, VkImageCreateInfo *ici, uint64_t mod)
482 {
483    if (check_ici(screen, ici, mod))
484       return true;
485 
486    ici->usage &= ~VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT;
487    if (check_ici(screen, ici, mod))
488       return true;
489 
490    ici->usage |= VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT;
491    return false;
492 }
493 
494 /* If the driver can't do mutable with this ICI, then try again after removing mutable (and
495  * thus also the list of formats we might might mutate to)
496  */
497 static bool
double_check_ici(struct zink_screen * screen,VkImageCreateInfo * ici,VkImageUsageFlags usage,uint64_t mod,bool require_mutable)498 double_check_ici(struct zink_screen *screen, VkImageCreateInfo *ici, VkImageUsageFlags usage, uint64_t mod, bool require_mutable)
499 {
500    if (!usage)
501       return false;
502 
503    ici->usage = usage;
504 
505    if (ici->usage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT) {
506       if (suboptimal_check_ici(screen, ici, mod))
507          return true;
508    }
509    if (check_ici(screen, ici, mod))
510       return true;
511    if (require_mutable)
512       return false;
513    const void *pNext = ici->pNext;
514    if (pNext) {
515       VkBaseOutStructure *prev = NULL;
516       VkBaseOutStructure *fmt_list = NULL;
517       vk_foreach_struct(strct, (void*)ici->pNext) {
518          if (strct->sType == VK_STRUCTURE_TYPE_IMAGE_FORMAT_LIST_CREATE_INFO) {
519             fmt_list = strct;
520             if (prev) {
521                prev->pNext = strct->pNext;
522             } else {
523                ici->pNext = strct->pNext;
524             }
525             fmt_list->pNext = NULL;
526             break;
527          }
528          prev = strct;
529       }
530       if (!fmt_list)
531          return false;
532       ici->flags &= ~VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
533       if (suboptimal_check_ici(screen, ici, mod))
534          return true;
535       fmt_list->pNext = (void*)ici->pNext;
536       ici->pNext = fmt_list;
537       ici->flags |= VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
538    }
539    return false;
540 }
541 
542 static bool
find_good_mod(struct zink_screen * screen,VkImageCreateInfo * ici,const struct pipe_resource * templ,unsigned bind,unsigned modifiers_count,uint64_t * modifiers,uint64_t * good_mod)543 find_good_mod(struct zink_screen *screen, VkImageCreateInfo *ici, const struct pipe_resource *templ, unsigned bind, unsigned modifiers_count, uint64_t *modifiers, uint64_t *good_mod)
544 {
545    bool found = false;
546    VkImageUsageFlags good_usage = 0;
547    const struct zink_modifier_props *prop = zink_get_modifier_props(screen, templ->format);
548    for (unsigned i = 0; i < modifiers_count; i++) {
549       bool need_extended = false;
550 
551       if (modifiers[i] == DRM_FORMAT_MOD_LINEAR)
552          continue;
553 
554       VkFormatFeatureFlags feats = find_modifier_feats(prop, modifiers[i]);
555       if (!feats)
556          continue;
557 
558       if (feats & VK_FORMAT_FEATURE_DISJOINT_BIT && util_format_get_num_planes(templ->format))
559          ici->flags |= VK_IMAGE_CREATE_DISJOINT_BIT;
560       VkImageUsageFlags usage = get_image_usage_for_feats(screen, feats, templ, bind, &need_extended);
561       assert(!need_extended);
562       if (double_check_ici(screen, ici, usage, modifiers[i], true)) {
563          /* assume "best" modifiers are last in array; just return last good modifier */
564          found = true;
565          *good_mod = modifiers[i];
566          good_usage = usage;
567       }
568    }
569    if (found)
570       ici->usage = good_usage;
571    return found;
572 }
573 
574 /* subfunctions of this call must set ici->usage on success */
575 static bool
set_image_usage(struct zink_screen * screen,VkImageCreateInfo * ici,const struct pipe_resource * templ,unsigned bind,unsigned modifiers_count,uint64_t * modifiers,uint64_t * mod)576 set_image_usage(struct zink_screen *screen, VkImageCreateInfo *ici, const struct pipe_resource *templ, unsigned bind, unsigned modifiers_count, uint64_t *modifiers, uint64_t *mod)
577 {
578    VkImageTiling tiling = ici->tiling;
579    bool need_extended = false;
580    *mod = DRM_FORMAT_MOD_INVALID;
581    if (modifiers_count) {
582       assert(tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT);
583       uint64_t good_mod = 0;
584       if (screen->info.have_EXT_image_drm_format_modifier &&
585           find_good_mod(screen, ici, templ, bind, modifiers_count, modifiers, &good_mod)) {
586          *mod = good_mod;
587          return true;
588       }
589       /* only try linear if no other options available */
590       const struct zink_modifier_props *prop = zink_get_modifier_props(screen, templ->format);
591       VkFormatFeatureFlags feats = find_modifier_feats(prop, DRM_FORMAT_MOD_LINEAR);
592       if (feats) {
593          if (feats & VK_FORMAT_FEATURE_DISJOINT_BIT && util_format_get_num_planes(templ->format) > 1)
594             ici->flags |= VK_IMAGE_CREATE_DISJOINT_BIT;
595          VkImageUsageFlags usage = get_image_usage_for_feats(screen, feats, templ, bind, &need_extended);
596          assert(!need_extended);
597          if (double_check_ici(screen, ici, usage, DRM_FORMAT_MOD_LINEAR, true)) {
598             *mod = DRM_FORMAT_MOD_LINEAR;
599             return true;
600          }
601       }
602    } else {
603       const struct zink_format_props *props = zink_get_format_props(screen, templ->format);
604       VkFormatFeatureFlags2 feats = tiling == VK_IMAGE_TILING_LINEAR ? props->linearTilingFeatures : props->optimalTilingFeatures;
605       if (feats & VK_FORMAT_FEATURE_DISJOINT_BIT && util_format_get_num_planes(templ->format) > 1)
606          ici->flags |= VK_IMAGE_CREATE_DISJOINT_BIT;
607       if (ici->flags & VK_IMAGE_CREATE_EXTENDED_USAGE_BIT)
608          feats = UINT32_MAX;
609       VkImageUsageFlags usage = get_image_usage_for_feats(screen, feats, templ, bind, &need_extended);
610       if (need_extended) {
611          ici->flags |= VK_IMAGE_CREATE_EXTENDED_USAGE_BIT | VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
612          feats = UINT32_MAX;
613          usage = get_image_usage_for_feats(screen, feats, templ, bind, &need_extended);
614       }
615       if (double_check_ici(screen, ici, usage, DRM_FORMAT_MOD_INVALID, true))
616          return true;
617       if (util_format_is_depth_or_stencil(templ->format)) {
618          if (!(templ->bind & PIPE_BIND_DEPTH_STENCIL)) {
619             usage &= ~VK_IMAGE_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT;
620             /* mutable doesn't apply to depth/stencil formats */
621             if (double_check_ici(screen, ici, usage, DRM_FORMAT_MOD_INVALID, true))
622                return true;
623          }
624       } else if (!(templ->bind & PIPE_BIND_RENDER_TARGET)) {
625          usage &= ~VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
626          if (double_check_ici(screen, ici, usage, DRM_FORMAT_MOD_INVALID, true))
627             return true;
628          usage |= VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
629          if (double_check_ici(screen, ici, usage, DRM_FORMAT_MOD_INVALID, false))
630             return true;
631          usage &= ~VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT;
632          if (double_check_ici(screen, ici, usage, DRM_FORMAT_MOD_INVALID, false))
633             return true;
634       } else {
635          if (double_check_ici(screen, ici, usage, DRM_FORMAT_MOD_INVALID, false))
636             return true;
637       }
638    }
639    ici->usage = 0;
640    *mod = DRM_FORMAT_MOD_INVALID;
641    return false;
642 }
643 
644 static bool
try_set_image_usage_or_EXTENDED(struct zink_screen * screen,VkImageCreateInfo * ici,const struct pipe_resource * templ,unsigned bind,unsigned modifiers_count,uint64_t * modifiers,uint64_t * mod)645 try_set_image_usage_or_EXTENDED(struct zink_screen *screen, VkImageCreateInfo *ici, const struct pipe_resource *templ, unsigned bind, unsigned modifiers_count, uint64_t *modifiers, uint64_t *mod)
646 {
647    VkImageCreateFlags flags = ici->flags;
648    /* retry with EXTENDED: trust that the frontend isn't giving us anything insane and pray */
649    ici->flags |= VK_IMAGE_CREATE_EXTENDED_USAGE_BIT | VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
650    bool success = set_image_usage(screen, ici, templ, bind, modifiers_count, modifiers, mod);
651    if (!success)
652       ici->flags = flags;
653    return success;
654 }
655 
656 static uint64_t
eval_ici(struct zink_screen * screen,VkImageCreateInfo * ici,const struct pipe_resource * templ,unsigned bind,unsigned modifiers_count,uint64_t * modifiers,bool * success)657 eval_ici(struct zink_screen *screen, VkImageCreateInfo *ici, const struct pipe_resource *templ, unsigned bind, unsigned modifiers_count, uint64_t *modifiers, bool *success)
658 {
659    /* sampleCounts will be set to VK_SAMPLE_COUNT_1_BIT if at least one of the following conditions is true:
660     * - flags contains VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT
661     *
662     * 44.1.1. Supported Sample Counts
663     */
664    bool want_cube = ici->samples == 1 &&
665                     (templ->target == PIPE_TEXTURE_CUBE ||
666                     templ->target == PIPE_TEXTURE_CUBE_ARRAY ||
667                     (templ->target == PIPE_TEXTURE_2D_ARRAY && ici->extent.width == ici->extent.height && ici->arrayLayers >= 6));
668 
669    if (ici->tiling != VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT)
670       modifiers_count = 0;
671 
672    uint64_t mod = DRM_FORMAT_MOD_INVALID;
673    /* this should work most of the time */
674    if (!set_image_usage(screen, ici, templ, bind, modifiers_count, modifiers, &mod)) {
675       bool success = false;
676       /* dmabuf doesn't retry with EXTENDED */
677       if (ici->tiling == VK_IMAGE_TILING_OPTIMAL)
678          success = try_set_image_usage_or_EXTENDED(screen, ici, templ, bind, modifiers_count, modifiers, &mod);
679       if (!success) {
680          ici->tiling = VK_IMAGE_TILING_LINEAR;
681          modifiers_count = 0;
682          /* in theory LINEAR should work */
683          if (!set_image_usage(screen, ici, templ, bind, modifiers_count, modifiers, &mod)) {
684             /* ...in theory LINEAR + EXTENDED should definitely work */
685             if (!try_set_image_usage_or_EXTENDED(screen, ici, templ, bind, modifiers_count, modifiers, &mod))
686                /* ...so that was a lie */
687                return DRM_FORMAT_MOD_INVALID;
688          }
689       }
690    }
691    if (want_cube) {
692       ici->flags |= VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT;
693       VkImageUsageFlags usage = ici->usage;
694       if (!set_image_usage(screen, ici, templ, bind, modifiers_count, modifiers, &mod)) {
695          ici->flags &= ~VK_IMAGE_CREATE_CUBE_COMPATIBLE_BIT;
696          ici->usage = usage;
697       }
698    }
699 
700    *success = true;
701    return mod;
702 }
703 
704 static void
init_ici(struct zink_screen * screen,VkImageCreateInfo * ici,const struct pipe_resource * templ,unsigned bind,unsigned modifiers_count)705 init_ici(struct zink_screen *screen, VkImageCreateInfo *ici, const struct pipe_resource *templ, unsigned bind, unsigned modifiers_count)
706 {
707    ici->sType = VK_STRUCTURE_TYPE_IMAGE_CREATE_INFO;
708    ici->format = zink_get_format(screen, templ->format);
709    ici->extent.width = templ->width0;
710    ici->extent.height = templ->height0;
711    ici->extent.depth = templ->depth0;
712    ici->mipLevels = templ->last_level + 1;
713    ici->arrayLayers = MAX2(templ->array_size, 1);
714    ici->samples = templ->nr_samples ? templ->nr_samples : VK_SAMPLE_COUNT_1_BIT;
715 
716    /* pNext may already be set */
717    if (bind & ZINK_BIND_MUTABLE)
718       ici->flags = VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
719    else
720       ici->flags = 0;
721    if (ici->flags & VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT)
722       /* unset VkImageFormatListCreateInfo if mutable */
723       ici->pNext = NULL;
724    else if (ici->pNext)
725       /* add mutable if VkImageFormatListCreateInfo */
726       ici->flags |= VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
727    ici->usage = 0;
728    ici->queueFamilyIndexCount = 0;
729    ici->pQueueFamilyIndices = NULL;
730 
731    /* assume we're going to be doing some CompressedTexSubImage */
732    if (util_format_is_compressed(templ->format) && (ici->flags & VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT) &&
733        !vk_find_struct_const(ici->pNext, IMAGE_FORMAT_LIST_CREATE_INFO))
734       ici->flags |= VK_IMAGE_CREATE_BLOCK_TEXEL_VIEW_COMPATIBLE_BIT;
735 
736    if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE)
737       ici->flags |= VK_IMAGE_CREATE_SPARSE_BINDING_BIT | VK_IMAGE_CREATE_SPARSE_RESIDENCY_BIT;
738 
739    bool need_2D = false;
740    switch (templ->target) {
741    case PIPE_TEXTURE_1D:
742    case PIPE_TEXTURE_1D_ARRAY:
743       if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE)
744          need_2D |= screen->need_2D_sparse;
745       if (util_format_is_depth_or_stencil(templ->format))
746          need_2D |= screen->need_2D_zs;
747       ici->imageType = need_2D ? VK_IMAGE_TYPE_2D : VK_IMAGE_TYPE_1D;
748       break;
749 
750    case PIPE_TEXTURE_CUBE:
751    case PIPE_TEXTURE_CUBE_ARRAY:
752    case PIPE_TEXTURE_2D:
753    case PIPE_TEXTURE_2D_ARRAY:
754    case PIPE_TEXTURE_RECT:
755       ici->imageType = VK_IMAGE_TYPE_2D;
756       break;
757 
758    case PIPE_TEXTURE_3D:
759       ici->imageType = VK_IMAGE_TYPE_3D;
760       if (!(templ->flags & PIPE_RESOURCE_FLAG_SPARSE)) {
761          ici->flags |= VK_IMAGE_CREATE_2D_ARRAY_COMPATIBLE_BIT;
762          if (screen->info.have_EXT_image_2d_view_of_3d)
763             ici->flags |= VK_IMAGE_CREATE_2D_VIEW_COMPATIBLE_BIT_EXT;
764       }
765       break;
766 
767    case PIPE_BUFFER:
768       unreachable("PIPE_BUFFER should already be handled");
769 
770    default:
771       unreachable("Unknown target");
772    }
773 
774    if (screen->info.have_EXT_sample_locations &&
775        bind & PIPE_BIND_DEPTH_STENCIL &&
776        util_format_has_depth(util_format_description(templ->format)))
777       ici->flags |= VK_IMAGE_CREATE_SAMPLE_LOCATIONS_COMPATIBLE_DEPTH_BIT_EXT;
778 
779    if (screen->info.have_EXT_image_drm_format_modifier && modifiers_count)
780       ici->tiling = VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT;
781    else if (bind & (PIPE_BIND_LINEAR | ZINK_BIND_DMABUF))
782       ici->tiling = VK_IMAGE_TILING_LINEAR;
783    else
784       ici->tiling = VK_IMAGE_TILING_OPTIMAL;
785    /* XXX: does this have perf implications anywhere? hopefully not */
786    if (ici->samples == VK_SAMPLE_COUNT_1_BIT &&
787       screen->info.have_EXT_multisampled_render_to_single_sampled &&
788       ici->tiling != VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT)
789       ici->flags |= VK_IMAGE_CREATE_MULTISAMPLED_RENDER_TO_SINGLE_SAMPLED_BIT_EXT;
790    ici->sharingMode = VK_SHARING_MODE_EXCLUSIVE;
791    ici->initialLayout = VK_IMAGE_LAYOUT_UNDEFINED;
792 
793    if (templ->target == PIPE_TEXTURE_CUBE)
794       ici->arrayLayers *= 6;
795 }
796 
797 static const VkImageAspectFlags plane_aspects[] = {
798    VK_IMAGE_ASPECT_PLANE_0_BIT,
799    VK_IMAGE_ASPECT_PLANE_1_BIT,
800    VK_IMAGE_ASPECT_PLANE_2_BIT,
801 };
802 
803 static inline bool
get_image_memory_requirement(struct zink_screen * screen,struct zink_resource_object * obj,unsigned num_planes,VkMemoryRequirements * reqs)804 get_image_memory_requirement(struct zink_screen *screen, struct zink_resource_object *obj,
805                              unsigned num_planes, VkMemoryRequirements *reqs)
806 {
807    bool need_dedicated = false;
808    if (VKSCR(GetImageMemoryRequirements2)) {
809       VkMemoryRequirements2 req2;
810       req2.sType = VK_STRUCTURE_TYPE_MEMORY_REQUIREMENTS_2;
811       VkImageMemoryRequirementsInfo2 info2;
812       info2.sType = VK_STRUCTURE_TYPE_IMAGE_MEMORY_REQUIREMENTS_INFO_2;
813       info2.pNext = NULL;
814       info2.image = obj->image;
815       VkMemoryDedicatedRequirements ded;
816       ded.sType = VK_STRUCTURE_TYPE_MEMORY_DEDICATED_REQUIREMENTS;
817       ded.pNext = NULL;
818       req2.pNext = &ded;
819       VkImagePlaneMemoryRequirementsInfo plane;
820       plane.sType = VK_STRUCTURE_TYPE_IMAGE_PLANE_MEMORY_REQUIREMENTS_INFO;
821       plane.pNext = NULL;
822       if (num_planes > 1)
823          info2.pNext = &plane;
824       unsigned offset = 0;
825       for (unsigned i = 0; i < num_planes; i++) {
826          assert(i < ARRAY_SIZE(plane_aspects));
827          plane.planeAspect = plane_aspects[i];
828          VKSCR(GetImageMemoryRequirements2)(screen->dev, &info2, &req2);
829          if (!i)
830             reqs->alignment = req2.memoryRequirements.alignment;
831          obj->plane_offsets[i] = offset;
832          offset += req2.memoryRequirements.size;
833          reqs->size += req2.memoryRequirements.size;
834          reqs->memoryTypeBits |= req2.memoryRequirements.memoryTypeBits;
835          need_dedicated |= ded.prefersDedicatedAllocation || ded.requiresDedicatedAllocation;
836       }
837    } else {
838       VKSCR(GetImageMemoryRequirements)(screen->dev, obj->image, reqs);
839    }
840    return need_dedicated;
841 }
842 
843 static inline VkFormatFeatureFlags
get_format_feature_flags(VkImageCreateInfo ici,struct zink_screen * screen,const struct pipe_resource * templ)844 get_format_feature_flags(VkImageCreateInfo ici, struct zink_screen *screen, const struct pipe_resource *templ)
845 {
846    VkFormatFeatureFlags feats = 0;
847    switch (ici.tiling) {
848    case VK_IMAGE_TILING_LINEAR:
849       feats = zink_get_format_props(screen, templ->format)->linearTilingFeatures;
850       break;
851    case VK_IMAGE_TILING_OPTIMAL:
852       feats = zink_get_format_props(screen, templ->format)->optimalTilingFeatures;
853       break;
854    case VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT:
855       feats = VK_FORMAT_FEATURE_FLAG_BITS_MAX_ENUM;
856       /*
857          If is tiling then VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT, the value of
858          imageCreateFormatFeatures is found by calling vkGetPhysicalDeviceFormatProperties2
859          with VkImageFormatProperties::format equal to VkImageCreateInfo::format and with
860          VkDrmFormatModifierPropertiesListEXT chained into VkImageFormatProperties2; by
861          collecting all members of the returned array
862          VkDrmFormatModifierPropertiesListEXT::pDrmFormatModifierProperties
863          whose drmFormatModifier belongs to imageCreateDrmFormatModifiers; and by taking the bitwise
864          intersection, over the collected array members, of drmFormatModifierTilingFeatures.
865          (The resultant imageCreateFormatFeatures may be empty).
866          * -Chapter 12. Resource Creation
867        */
868       for (unsigned i = 0; i < screen->modifier_props[templ->format].drmFormatModifierCount; i++)
869          feats &= screen->modifier_props[templ->format].pDrmFormatModifierProperties[i].drmFormatModifierTilingFeatures;
870       break;
871    default:
872          unreachable("unknown tiling");
873    }
874    return feats;
875 }
876 
877 #if !defined(_WIN32)
878    #define ZINK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_BIT VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_FD_BIT
879 #else
880    #define ZINK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_BIT VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_WIN32_BIT
881 #endif
882 
883 
884 struct mem_alloc_info {
885    struct winsys_handle *whandle;
886    VkMemoryPropertyFlags flags;
887    enum zink_alloc_flag aflags;
888    bool need_dedicated;
889    bool shared;
890    const void *user_mem;
891    VkExternalMemoryHandleTypeFlags external;
892    VkExternalMemoryHandleTypeFlags export_types;
893 };
894 
895 static inline bool
get_export_flags(struct zink_screen * screen,const struct pipe_resource * templ,struct mem_alloc_info * alloc_info)896 get_export_flags(struct zink_screen *screen, const struct pipe_resource *templ, struct mem_alloc_info *alloc_info)
897 {
898    bool needs_export = (templ->bind & (ZINK_BIND_VIDEO | ZINK_BIND_DMABUF)) != 0;
899    if (alloc_info->whandle) {
900       if (alloc_info->whandle->type == WINSYS_HANDLE_TYPE_FD ||
901           alloc_info->whandle->type == ZINK_EXTERNAL_MEMORY_HANDLE)
902          needs_export |= true;
903       else
904          unreachable("unknown handle type");
905    }
906    if (needs_export) {
907       if (alloc_info->whandle && alloc_info->whandle->type == ZINK_EXTERNAL_MEMORY_HANDLE) {
908          alloc_info->external = ZINK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_BIT;
909       } else if (screen->info.have_EXT_external_memory_dma_buf) {
910          alloc_info->external = VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
911          alloc_info->export_types |= VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
912       } else {
913          return false;
914       }
915    }
916    if (alloc_info->user_mem) {
917       assert(!alloc_info->whandle);
918       alloc_info->external = VK_EXTERNAL_MEMORY_HANDLE_TYPE_HOST_ALLOCATION_BIT_EXT;
919       alloc_info->export_types = VK_EXTERNAL_MEMORY_HANDLE_TYPE_HOST_ALLOCATION_BIT_EXT;
920    }
921    /* we may export WINSYS_HANDLE_TYPE_FD handle which is dma-buf */
922    if (templ->bind & PIPE_BIND_SHARED && screen->info.have_EXT_external_memory_dma_buf)
923       alloc_info->export_types |= VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
924    return true;
925 }
926 
927 enum resource_object_create_result {
928    roc_success,
929    roc_success_early_return,
930    roc_fail_and_free_object,
931    roc_fail_and_cleanup_object,
932    roc_fail_and_cleanup_all
933 };
934 
935 static inline enum resource_object_create_result
allocate_bo(struct zink_screen * screen,const struct pipe_resource * templ,VkMemoryRequirements * reqs,struct zink_resource_object * obj,struct mem_alloc_info * alloc_info)936 allocate_bo(struct zink_screen *screen, const struct pipe_resource *templ,
937             VkMemoryRequirements *reqs, struct zink_resource_object *obj,
938             struct mem_alloc_info *alloc_info)
939 {
940    VkMemoryAllocateInfo mai;
941    mai.sType = VK_STRUCTURE_TYPE_MEMORY_ALLOCATE_INFO;
942    mai.pNext = NULL;
943    mai.allocationSize = reqs->size;
944    enum zink_heap heap = zink_heap_from_domain_flags(alloc_info->flags, alloc_info->aflags);
945    if (templ->flags & PIPE_RESOURCE_FLAG_MAP_COHERENT) {
946       if (!(vk_domain_from_heap(heap) & VK_MEMORY_PROPERTY_HOST_COHERENT_BIT))
947          heap = zink_heap_from_domain_flags(alloc_info->flags & ~VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT, alloc_info->aflags);
948    }
949 
950    VkMemoryDedicatedAllocateInfo ded_alloc_info = {
951       .sType = VK_STRUCTURE_TYPE_MEMORY_DEDICATED_ALLOCATE_INFO,
952       .pNext = mai.pNext,
953       .image = obj->image,
954       .buffer = VK_NULL_HANDLE,
955    };
956 
957    if (screen->info.have_KHR_dedicated_allocation && alloc_info->need_dedicated) {
958       ded_alloc_info.pNext = mai.pNext;
959       mai.pNext = &ded_alloc_info;
960    }
961 
962    VkExportMemoryAllocateInfo emai;
963    if ((templ->bind & ZINK_BIND_VIDEO) || ((templ->bind & PIPE_BIND_SHARED) && alloc_info->shared) || (templ->bind & ZINK_BIND_DMABUF)) {
964       emai.sType = VK_STRUCTURE_TYPE_EXPORT_MEMORY_ALLOCATE_INFO;
965       emai.handleTypes = alloc_info->export_types;
966 
967       emai.pNext = mai.pNext;
968       mai.pNext = &emai;
969       obj->exportable = true;
970    }
971 
972 #ifdef ZINK_USE_DMABUF
973 
974 #if !defined(_WIN32)
975    VkImportMemoryFdInfoKHR imfi = {
976       VK_STRUCTURE_TYPE_IMPORT_MEMORY_FD_INFO_KHR,
977       NULL,
978    };
979 
980    if (alloc_info->whandle) {
981       imfi.pNext = NULL;
982       imfi.handleType = alloc_info->external;
983       imfi.fd = os_dupfd_cloexec(alloc_info->whandle->handle);
984       if (imfi.fd < 0) {
985          mesa_loge("ZINK: failed to dup dmabuf fd: %s\n", strerror(errno));
986          return roc_fail_and_cleanup_object;
987       }
988 
989       imfi.pNext = mai.pNext;
990       mai.pNext = &imfi;
991    }
992 #else
993    VkImportMemoryWin32HandleInfoKHR imfi = {
994       VK_STRUCTURE_TYPE_IMPORT_MEMORY_WIN32_HANDLE_INFO_KHR,
995       NULL,
996    };
997 
998    if (alloc_info->whandle) {
999       HANDLE source_target = GetCurrentProcess();
1000       HANDLE out_handle;
1001 
1002       bool result = DuplicateHandle(source_target, alloc_info->whandle->handle, source_target, &out_handle, 0, false, DUPLICATE_SAME_ACCESS);
1003 
1004       if (!result || !out_handle) {
1005          mesa_loge("ZINK: failed to DuplicateHandle with winerr: %08x\n", (int)GetLastError());
1006          return roc_fail_and_cleanup_object;
1007       }
1008 
1009       imfi.pNext = NULL;
1010       imfi.handleType = alloc_info->external;
1011       imfi.handle = out_handle;
1012 
1013       imfi.pNext = mai.pNext;
1014       mai.pNext = &imfi;
1015    }
1016 #endif
1017 
1018 #endif
1019 
1020    VkImportMemoryHostPointerInfoEXT imhpi = {
1021       VK_STRUCTURE_TYPE_IMPORT_MEMORY_HOST_POINTER_INFO_EXT,
1022       NULL,
1023    };
1024    if (alloc_info->user_mem) {
1025       imhpi.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_HOST_ALLOCATION_BIT_EXT;
1026       imhpi.pHostPointer = (void*)alloc_info->user_mem;
1027       imhpi.pNext = mai.pNext;
1028       mai.pNext = &imhpi;
1029    }
1030 
1031    unsigned alignment = MAX2(reqs->alignment, 256);
1032    if (templ->usage == PIPE_USAGE_STAGING && obj->is_buffer)
1033       alignment = MAX2(alignment, screen->info.props.limits.minMemoryMapAlignment);
1034    obj->alignment = alignment;
1035 
1036    if (zink_mem_type_idx_from_types(screen, heap, reqs->memoryTypeBits) == UINT32_MAX) {
1037       /* not valid based on reqs; demote to more compatible type */
1038       switch (heap) {
1039       case ZINK_HEAP_DEVICE_LOCAL_VISIBLE:
1040          heap = ZINK_HEAP_DEVICE_LOCAL;
1041          break;
1042       case ZINK_HEAP_HOST_VISIBLE_COHERENT_CACHED:
1043          heap = ZINK_HEAP_HOST_VISIBLE_COHERENT;
1044          break;
1045       default:
1046          break;
1047       }
1048       assert(zink_mem_type_idx_from_types(screen, heap, reqs->memoryTypeBits) != UINT32_MAX);
1049    }
1050 
1051    while (1) {
1052       /* iterate over all available memory types to reduce chance of oom */
1053       for (unsigned i = 0; !obj->bo && i < screen->heap_count[heap]; i++) {
1054          if (!(reqs->memoryTypeBits & BITFIELD_BIT(screen->heap_map[heap][i])))
1055             continue;
1056 
1057          mai.memoryTypeIndex = screen->heap_map[heap][i];
1058          obj->bo = zink_bo(zink_bo_create(screen, reqs->size, alignment, heap, mai.pNext ? ZINK_ALLOC_NO_SUBALLOC : 0, mai.memoryTypeIndex, mai.pNext));
1059       }
1060 
1061       if (obj->bo || heap != ZINK_HEAP_DEVICE_LOCAL_VISIBLE)
1062          break;
1063 
1064       /* demote BAR allocations to a different heap on failure to avoid oom */
1065       if (templ->flags & PIPE_RESOURCE_FLAG_MAP_COHERENT || templ->usage == PIPE_USAGE_DYNAMIC)
1066           heap = ZINK_HEAP_HOST_VISIBLE_COHERENT;
1067       else
1068           heap = ZINK_HEAP_DEVICE_LOCAL;
1069    };
1070 
1071    return obj->bo ? roc_success : roc_fail_and_cleanup_object;
1072 }
1073 
1074 static inline bool
update_alloc_info_flags(struct zink_screen * screen,const struct pipe_resource * templ,VkMemoryRequirements * reqs,struct mem_alloc_info * alloc_info)1075 update_alloc_info_flags(struct zink_screen *screen, const struct pipe_resource *templ,
1076                         VkMemoryRequirements *reqs, struct mem_alloc_info *alloc_info)
1077 {
1078    if (templ->flags & PIPE_RESOURCE_FLAG_MAP_COHERENT || templ->usage == PIPE_USAGE_DYNAMIC)
1079       alloc_info->flags |= VK_MEMORY_PROPERTY_HOST_COHERENT_BIT;
1080    else if (!(alloc_info->flags & VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT) &&
1081             templ->usage == PIPE_USAGE_STAGING)
1082       alloc_info->flags |= VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT;
1083 
1084    if (templ->bind & ZINK_BIND_TRANSIENT)
1085       alloc_info->flags |= VK_MEMORY_PROPERTY_LAZILY_ALLOCATED_BIT;
1086 
1087    if (alloc_info->user_mem) {
1088       VkExternalMemoryHandleTypeFlagBits handle_type = VK_EXTERNAL_MEMORY_HANDLE_TYPE_HOST_ALLOCATION_BIT_EXT;
1089       VkMemoryHostPointerPropertiesEXT memory_host_pointer_properties = {0};
1090       memory_host_pointer_properties.sType = VK_STRUCTURE_TYPE_MEMORY_HOST_POINTER_PROPERTIES_EXT;
1091       memory_host_pointer_properties.pNext = NULL;
1092       VkResult res = VKSCR(GetMemoryHostPointerPropertiesEXT)(screen->dev, handle_type, alloc_info->user_mem, &memory_host_pointer_properties);
1093       if (res != VK_SUCCESS) {
1094          mesa_loge("ZINK: vkGetMemoryHostPointerPropertiesEXT failed");
1095          return false;
1096       }
1097       reqs->memoryTypeBits &= memory_host_pointer_properties.memoryTypeBits;
1098       alloc_info->flags &= ~VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
1099    }
1100 
1101    alloc_info->aflags = templ->flags & PIPE_RESOURCE_FLAG_SPARSE ? ZINK_ALLOC_SPARSE : 0;
1102    return true;
1103 }
1104 
1105 static inline void
update_obj_info(struct zink_screen * screen,struct zink_resource_object * obj,const struct pipe_resource * templ,struct mem_alloc_info * alloc_info)1106 update_obj_info(struct zink_screen *screen, struct zink_resource_object *obj,
1107                 const struct pipe_resource *templ, struct mem_alloc_info *alloc_info)
1108 {
1109    if (alloc_info->aflags == ZINK_ALLOC_SPARSE) {
1110       obj->size = templ->width0;
1111    } else {
1112       obj->offset = zink_bo_get_offset(obj->bo);
1113       obj->size = zink_bo_get_size(obj->bo);
1114    }
1115 
1116    obj->coherent = screen->info.mem_props.memoryTypes[obj->bo->base.base.placement].propertyFlags & VK_MEMORY_PROPERTY_HOST_COHERENT_BIT;
1117    if (!(templ->flags & PIPE_RESOURCE_FLAG_SPARSE)) {
1118       obj->host_visible = screen->info.mem_props.memoryTypes[obj->bo->base.base.placement].propertyFlags & VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
1119    }
1120 }
1121 
1122 static inline void
debug_resource_mem(struct zink_resource_object * obj,const struct pipe_resource * templ,struct zink_screen * screen)1123 debug_resource_mem(struct zink_resource_object *obj, const struct pipe_resource *templ, struct zink_screen *screen)
1124 {
1125       char buf[4096];
1126       unsigned idx = 0;
1127       if (obj->is_buffer) {
1128          size_t size = (size_t)DIV_ROUND_UP(obj->size, 1024);
1129          if (templ->bind == PIPE_BIND_QUERY_BUFFER && templ->usage == PIPE_USAGE_STAGING) //internal qbo
1130             idx += snprintf(buf, sizeof(buf), "QBO(%zu)", size);
1131          else
1132             idx += snprintf(buf, sizeof(buf), "BUF(%zu)", size);
1133       } else {
1134          idx += snprintf(buf, sizeof(buf), "IMG(%s:%ux%ux%u)", util_format_short_name(templ->format), templ->width0, templ->height0, templ->depth0);
1135       }
1136       /*
1137       zink_vkflags_func flag_func = obj->is_buffer ? (zink_vkflags_func)vk_BufferCreateFlagBits_to_str : (zink_vkflags_func)vk_ImageCreateFlagBits_to_str;
1138       zink_vkflags_func usage_func = obj->is_buffer ? (zink_vkflags_func)vk_BufferUsageFlagBits_to_str : (zink_vkflags_func)vk_ImageUsageFlagBits_to_str;
1139       if (obj->vkflags) {
1140          buf[idx++] = '[';
1141          idx += zink_string_vkflags_unroll(&buf[idx], sizeof(buf) - idx, obj->vkflags, flag_func);
1142          buf[idx++] = ']';
1143       }
1144       if (obj->vkusage) {
1145          buf[idx++] = '[';
1146          idx += zink_string_vkflags_unroll(&buf[idx], sizeof(buf) - idx, obj->vkusage, usage_func);
1147          buf[idx++] = ']';
1148       }
1149       */
1150       buf[idx] = 0;
1151       obj->bo->name = zink_debug_mem_add(screen, obj->size, buf);
1152 }
1153 
1154 static inline enum resource_object_create_result
allocate_bo_and_update_obj(struct zink_screen * screen,const struct pipe_resource * templ,VkMemoryRequirements * reqs,struct zink_resource_object * obj,struct mem_alloc_info * alloc_info)1155 allocate_bo_and_update_obj(struct zink_screen *screen, const struct pipe_resource *templ,
1156                            VkMemoryRequirements *reqs, struct zink_resource_object *obj,
1157                            struct mem_alloc_info *alloc_info)
1158 {
1159    if (!update_alloc_info_flags(screen, templ, reqs, alloc_info))
1160       return roc_fail_and_free_object;
1161 
1162    enum resource_object_create_result retval = allocate_bo(screen, templ, reqs, obj, alloc_info);
1163    assert(retval != roc_success_early_return);
1164    if (retval != roc_success)
1165       return retval;
1166 
1167    update_obj_info(screen, obj, templ, alloc_info);
1168 
1169    if (zink_debug & ZINK_DEBUG_MEM)
1170       debug_resource_mem(obj, templ, screen);
1171    return roc_success;
1172 }
1173 
1174 static inline enum resource_object_create_result
create_buffer(struct zink_screen * screen,struct zink_resource_object * obj,const struct pipe_resource * templ,uint64_t * modifiers,int modifiers_count,struct mem_alloc_info * alloc_info)1175 create_buffer(struct zink_screen *screen, struct zink_resource_object *obj,
1176               const struct pipe_resource *templ, uint64_t *modifiers,
1177               int modifiers_count, struct mem_alloc_info *alloc_info)
1178 {
1179    VkBufferCreateInfo bci = create_bci(screen, templ, templ->bind);
1180    VkExternalMemoryBufferCreateInfo embci;
1181    VkMemoryRequirements reqs = {0};
1182 
1183    embci.sType = VK_STRUCTURE_TYPE_EXTERNAL_MEMORY_BUFFER_CREATE_INFO;
1184    if (alloc_info->external) {
1185       embci.pNext = bci.pNext;
1186       embci.handleTypes = alloc_info->export_types;
1187       bci.pNext = &embci;
1188    }
1189 
1190    if (VKSCR(CreateBuffer)(screen->dev, &bci, NULL, &obj->buffer) != VK_SUCCESS) {
1191       mesa_loge("ZINK: vkCreateBuffer failed");
1192       return roc_fail_and_free_object;
1193    }
1194 
1195    if (!(templ->bind & (PIPE_BIND_SHADER_IMAGE | ZINK_BIND_DESCRIPTOR))) {
1196        bci.usage |= VK_BUFFER_USAGE_STORAGE_TEXEL_BUFFER_BIT;
1197      if (VKSCR(CreateBuffer)(screen->dev, &bci, NULL, &obj->storage_buffer) != VK_SUCCESS) {
1198         mesa_loge("ZINK: vkCreateBuffer failed");
1199         VKSCR(DestroyBuffer)(screen->dev, obj->buffer, NULL);
1200         return roc_fail_and_free_object;
1201      }
1202    }
1203 
1204    if (modifiers_count) {
1205       assert(modifiers_count == 3);
1206       /* this is the DGC path because there's no other way to pass mem bits and I don't wanna copy/paste everything around */
1207       reqs.size = modifiers[0];
1208       reqs.alignment = modifiers[1];
1209       reqs.memoryTypeBits = modifiers[2];
1210    } else {
1211       VKSCR(GetBufferMemoryRequirements)(screen->dev, obj->buffer, &reqs);
1212    }
1213 
1214    if (templ->usage == PIPE_USAGE_STAGING)
1215       alloc_info->flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_HOST_COHERENT_BIT | VK_MEMORY_PROPERTY_HOST_CACHED_BIT;
1216    else if (templ->usage == PIPE_USAGE_STREAM)
1217       alloc_info->flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
1218    else if (templ->usage == PIPE_USAGE_IMMUTABLE)
1219       alloc_info->flags = VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
1220    else
1221       alloc_info->flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT | VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
1222 
1223    obj->is_buffer = true;
1224    obj->transfer_dst = true;
1225    obj->vkflags = bci.flags;
1226    obj->vkusage = bci.usage;
1227 
1228    enum resource_object_create_result retval = allocate_bo_and_update_obj(screen, templ, &reqs, obj,  alloc_info);
1229    assert(retval != roc_success_early_return);
1230    if (retval != roc_success)
1231       return retval;
1232 
1233    if (!(templ->flags & PIPE_RESOURCE_FLAG_SPARSE)) {
1234       if (VKSCR(BindBufferMemory)(screen->dev, obj->buffer, zink_bo_get_mem(obj->bo), obj->offset) != VK_SUCCESS) {
1235          mesa_loge("ZINK: vkBindBufferMemory failed");
1236          return roc_fail_and_cleanup_all ;
1237       }
1238       if (obj->storage_buffer && VKSCR(BindBufferMemory)(screen->dev, obj->storage_buffer, zink_bo_get_mem(obj->bo), obj->offset) != VK_SUCCESS) {
1239          mesa_loge("ZINK: vkBindBufferMemory failed");
1240          return roc_fail_and_cleanup_all;
1241       }
1242    }
1243    return roc_success;
1244 }
1245 
1246 static inline enum resource_object_create_result
create_image(struct zink_screen * screen,struct zink_resource_object * obj,const struct pipe_resource * templ,bool * linear,uint64_t * modifiers,int modifiers_count,struct mem_alloc_info * alloc_info)1247 create_image(struct zink_screen *screen, struct zink_resource_object *obj,
1248              const struct pipe_resource *templ, bool *linear,
1249              uint64_t *modifiers, int modifiers_count,
1250              struct mem_alloc_info *alloc_info)
1251 {
1252    VkMemoryRequirements reqs = {0};
1253    bool winsys_modifier = (alloc_info->export_types & VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT) &&
1254                           alloc_info->whandle &&
1255                           alloc_info->whandle->modifier != DRM_FORMAT_MOD_INVALID;
1256    uint64_t *ici_modifiers = winsys_modifier ? &alloc_info->whandle->modifier : modifiers;
1257    unsigned ici_modifier_count = winsys_modifier ? 1 : modifiers_count;
1258    unsigned num_planes = util_format_get_num_planes(templ->format);
1259    VkImageCreateInfo ici;
1260    enum pipe_format srgb = PIPE_FORMAT_NONE;
1261    /* we often need to be able to mutate between srgb and linear, but we don't need general
1262     * image view/shader image format compatibility (that path means losing fast clears or compression on some hardware).
1263     */
1264    if (!(templ->bind & ZINK_BIND_MUTABLE) && (!alloc_info->whandle || alloc_info->whandle->type == ZINK_EXTERNAL_MEMORY_HANDLE)) {
1265       srgb = util_format_is_srgb(templ->format) ? util_format_linear(templ->format) : util_format_srgb(templ->format);
1266       /* why do these helpers have different default return values? */
1267       if (srgb == templ->format)
1268          srgb = PIPE_FORMAT_NONE;
1269    }
1270    VkFormat formats[4] = {VK_FORMAT_UNDEFINED};
1271    VkImageFormatListCreateInfo format_list;
1272    if (srgb) {
1273       formats[0] = zink_get_format(screen, templ->format);
1274       formats[1] = zink_get_format(screen, srgb);
1275    } else if (templ->bind & ZINK_BIND_VIDEO) {
1276       formats[0] = zink_get_format(screen, templ->format);
1277       for (unsigned i = 0; i < num_planes; i++)
1278          formats[i + 1] = zink_get_format(screen, util_format_get_plane_format(templ->format, i));
1279    }
1280    /* only use format list if multiple formats have supported vk equivalents */
1281    if (formats[0] && formats[1]) {
1282       format_list.sType = VK_STRUCTURE_TYPE_IMAGE_FORMAT_LIST_CREATE_INFO;
1283       format_list.pNext = NULL;
1284       format_list.viewFormatCount = formats[2] ? 3 : 2;
1285       format_list.pViewFormats = formats;
1286       ici.pNext = &format_list;
1287    } else {
1288       ici.pNext = NULL;
1289    }
1290    init_ici(screen, &ici, templ, templ->bind, ici_modifier_count);
1291 
1292    bool success = false;
1293    uint64_t mod = eval_ici(screen, &ici, templ, templ->bind, ici_modifier_count, ici_modifiers, &success);
1294    if (ici.format == VK_FORMAT_A8_UNORM_KHR && !success) {
1295       ici.format = zink_get_format(screen, zink_format_get_emulated_alpha(templ->format));
1296       mod = eval_ici(screen, &ici, templ, templ->bind, ici_modifier_count, ici_modifiers, &success);
1297    }
1298    if (!success)
1299       return roc_fail_and_free_object;
1300 
1301    if (ici.tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT && srgb &&
1302       util_format_get_nr_components(srgb) == 4 &&
1303       !(ici.flags & VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT)) {
1304          mesa_loge("zink: refusing to create possibly-srgb dmabuf due to missing driver support: %s not supported!", util_format_name(srgb));
1305          return roc_fail_and_free_object;
1306    }
1307    VkExternalMemoryImageCreateInfo emici;
1308    VkImageDrmFormatModifierExplicitCreateInfoEXT idfmeci;
1309    VkImageDrmFormatModifierListCreateInfoEXT idfmlci;
1310    VkSubresourceLayout plane_layouts[4];
1311    VkSubresourceLayout plane_layout = {
1312       .offset = alloc_info->whandle ? alloc_info->whandle->offset : 0,
1313       .size = 0,
1314       .rowPitch = alloc_info->whandle ? alloc_info->whandle->stride : 0,
1315       .arrayPitch = 0,
1316       .depthPitch = 0,
1317    };
1318 
1319    obj->render_target = (ici.usage & VK_IMAGE_USAGE_COLOR_ATTACHMENT_BIT) != 0;
1320 
1321    if (ici.tiling == VK_IMAGE_TILING_OPTIMAL) {
1322       alloc_info->external &= ~VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
1323       alloc_info->export_types &= ~VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
1324    }
1325 
1326    if (alloc_info->shared || alloc_info->external) {
1327       emici.sType = VK_STRUCTURE_TYPE_EXTERNAL_MEMORY_IMAGE_CREATE_INFO;
1328       emici.pNext = ici.pNext;
1329       emici.handleTypes = alloc_info->export_types;
1330       assert(!(emici.handleTypes & VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT) || ici.tiling != VK_IMAGE_TILING_OPTIMAL);
1331       ici.pNext = &emici;
1332 
1333       assert(ici.tiling != VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT || mod != DRM_FORMAT_MOD_INVALID);
1334       if (alloc_info->whandle && ici.tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT) {
1335          assert(mod == alloc_info->whandle->modifier || !winsys_modifier);
1336          idfmeci.sType = VK_STRUCTURE_TYPE_IMAGE_DRM_FORMAT_MODIFIER_EXPLICIT_CREATE_INFO_EXT;
1337          idfmeci.pNext = ici.pNext;
1338          idfmeci.drmFormatModifier = mod;
1339          idfmeci.drmFormatModifierPlaneCount = obj->plane_count;
1340 
1341          plane_layouts[0] = plane_layout;
1342          struct pipe_resource *pnext = templ->next;
1343          for (unsigned i = 1; i < obj->plane_count; i++, pnext = pnext->next) {
1344             struct zink_resource *next = zink_resource(pnext);
1345             obj->plane_offsets[i] = plane_layouts[i].offset = next->obj->plane_offsets[i];
1346             obj->plane_strides[i] = plane_layouts[i].rowPitch = next->obj->plane_strides[i];
1347             plane_layouts[i].size = 0;
1348             plane_layouts[i].arrayPitch = 0;
1349             plane_layouts[i].depthPitch = 0;
1350          }
1351          idfmeci.pPlaneLayouts = plane_layouts;
1352 
1353          ici.pNext = &idfmeci;
1354       } else if (ici.tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT) {
1355          idfmlci.sType = VK_STRUCTURE_TYPE_IMAGE_DRM_FORMAT_MODIFIER_LIST_CREATE_INFO_EXT;
1356          idfmlci.pNext = ici.pNext;
1357          idfmlci.drmFormatModifierCount = modifiers_count;
1358          idfmlci.pDrmFormatModifiers = modifiers;
1359          ici.pNext = &idfmlci;
1360       } else if (ici.tiling == VK_IMAGE_TILING_OPTIMAL) {
1361          alloc_info->shared = false;
1362       }
1363    } else if (alloc_info->user_mem) {
1364       emici.sType = VK_STRUCTURE_TYPE_EXTERNAL_MEMORY_IMAGE_CREATE_INFO;
1365       emici.pNext = ici.pNext;
1366       emici.handleTypes = VK_EXTERNAL_MEMORY_HANDLE_TYPE_HOST_ALLOCATION_BIT_EXT;
1367       ici.pNext = &emici;
1368    }
1369 
1370    if (linear)
1371       *linear = ici.tiling == VK_IMAGE_TILING_LINEAR;
1372 
1373    if (ici.usage & VK_IMAGE_USAGE_TRANSFER_DST_BIT)
1374       obj->transfer_dst = true;
1375 
1376 #if defined(ZINK_USE_DMABUF) && !defined(_WIN32)
1377    if (obj->is_aux) {
1378       obj->modifier = mod;
1379       obj->modifier_aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_0_BIT_EXT << alloc_info->whandle->plane;
1380       obj->plane_offsets[alloc_info->whandle->plane] = alloc_info->whandle->offset;
1381       obj->plane_strides[alloc_info->whandle->plane] = alloc_info->whandle->stride;
1382       obj->handle = os_dupfd_cloexec(alloc_info->whandle->handle);
1383       if (obj->handle < 0) {
1384          mesa_loge("ZINK: failed to dup dmabuf fd: %s\n", strerror(errno));
1385          return roc_fail_and_free_object;
1386       }
1387       return roc_success_early_return;
1388    }
1389 #endif
1390 
1391    if (!(templ->bind & ZINK_BIND_VIDEO)) {
1392       obj->vkfeats = get_format_feature_flags(ici, screen, templ);
1393       if (obj->vkfeats & VK_FORMAT_FEATURE_DISJOINT_BIT)
1394          ici.flags |= VK_IMAGE_CREATE_DISJOINT_BIT;
1395    }
1396    if (util_format_is_yuv(templ->format)) {
1397    } else if (alloc_info->whandle) {
1398       obj->plane_strides[alloc_info->whandle->plane] = alloc_info->whandle->stride;
1399    }
1400 
1401    VkResult result = VKSCR(CreateImage)(screen->dev, &ici, NULL, &obj->image);
1402    if (result != VK_SUCCESS) {
1403       mesa_loge("ZINK: vkCreateImage failed (%s)", vk_Result_to_str(result));
1404       return roc_fail_and_free_object;
1405    }
1406 
1407    if (ici.tiling == VK_IMAGE_TILING_DRM_FORMAT_MODIFIER_EXT) {
1408       VkImageDrmFormatModifierPropertiesEXT modprops = {0};
1409       modprops.sType = VK_STRUCTURE_TYPE_IMAGE_DRM_FORMAT_MODIFIER_PROPERTIES_EXT;
1410       result = VKSCR(GetImageDrmFormatModifierPropertiesEXT)(screen->dev, obj->image, &modprops);
1411       if (result != VK_SUCCESS) {
1412          mesa_loge("ZINK: vkGetImageDrmFormatModifierPropertiesEXT failed");
1413          return roc_fail_and_free_object;
1414       }
1415       obj->modifier = modprops.drmFormatModifier;
1416       unsigned num_dmabuf_planes = screen->base.get_dmabuf_modifier_planes(&screen->base, obj->modifier, templ->format);
1417       obj->modifier_aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_0_BIT_EXT;
1418       if (num_dmabuf_planes > 1)
1419          obj->modifier_aspect |= VK_IMAGE_ASPECT_MEMORY_PLANE_1_BIT_EXT;
1420       if (num_dmabuf_planes > 2)
1421          obj->modifier_aspect |= VK_IMAGE_ASPECT_MEMORY_PLANE_2_BIT_EXT;
1422       if (num_dmabuf_planes > 3)
1423          obj->modifier_aspect |= VK_IMAGE_ASPECT_MEMORY_PLANE_3_BIT_EXT;
1424       assert(num_dmabuf_planes <= 4);
1425    }
1426 
1427    alloc_info->need_dedicated = get_image_memory_requirement(screen, obj, num_planes, &reqs);
1428    if (templ->usage == PIPE_USAGE_STAGING && ici.tiling == VK_IMAGE_TILING_LINEAR)
1429       alloc_info->flags = VK_MEMORY_PROPERTY_HOST_VISIBLE_BIT;
1430    else
1431       alloc_info->flags = VK_MEMORY_PROPERTY_DEVICE_LOCAL_BIT;
1432 
1433    obj->vkflags = ici.flags;
1434    obj->vkusage = ici.usage;
1435 
1436    enum resource_object_create_result retval = allocate_bo_and_update_obj(screen, templ, &reqs, obj,  alloc_info);
1437    assert(retval != roc_success_early_return);
1438    if (retval != roc_success)
1439       return retval;
1440 
1441    if (ici.flags & VK_IMAGE_CREATE_DISJOINT_BIT) {
1442       VkBindImageMemoryInfo infos[3];
1443       VkBindImagePlaneMemoryInfo planes[3];
1444       for (unsigned i = 0; i < num_planes; i++) {
1445          infos[i].sType = VK_STRUCTURE_TYPE_BIND_IMAGE_MEMORY_INFO;
1446          infos[i].image = obj->image;
1447          infos[i].memory = zink_bo_get_mem(obj->bo);
1448          infos[i].memoryOffset = obj->plane_offsets[i];
1449          if (templ->bind & ZINK_BIND_VIDEO) {
1450             infos[i].pNext = &planes[i];
1451             planes[i].sType = VK_STRUCTURE_TYPE_BIND_IMAGE_PLANE_MEMORY_INFO;
1452             planes[i].pNext = NULL;
1453             planes[i].planeAspect = plane_aspects[i];
1454          }
1455       }
1456       if (VKSCR(BindImageMemory2)(screen->dev, num_planes, infos) != VK_SUCCESS) {
1457          mesa_loge("ZINK: vkBindImageMemory2 failed");
1458          return roc_fail_and_cleanup_all;
1459       }
1460    } else {
1461       if (!(templ->flags & PIPE_RESOURCE_FLAG_SPARSE))
1462          if (VKSCR(BindImageMemory)(screen->dev, obj->image, zink_bo_get_mem(obj->bo), obj->offset) != VK_SUCCESS) {
1463             mesa_loge("ZINK: vkBindImageMemory failed");
1464             return roc_fail_and_cleanup_all;
1465          }
1466    }
1467 
1468    return roc_success;
1469 }
1470 
1471 static struct zink_resource_object *
resource_object_create(struct zink_screen * screen,const struct pipe_resource * templ,struct winsys_handle * whandle,bool * linear,uint64_t * modifiers,int modifiers_count,const void * loader_private,const void * user_mem)1472 resource_object_create(struct zink_screen *screen, const struct pipe_resource *templ, struct winsys_handle *whandle, bool *linear,
1473                        uint64_t *modifiers, int modifiers_count, const void *loader_private, const void *user_mem)
1474 {
1475    struct zink_resource_object *obj = CALLOC_STRUCT(zink_resource_object);
1476    unsigned max_level = 0;
1477    if (!obj)
1478       return NULL;
1479    simple_mtx_init(&obj->view_lock, mtx_plain);
1480    util_dynarray_init(&obj->views, NULL);
1481    u_rwlock_init(&obj->copy_lock);
1482    obj->unordered_read = true;
1483    obj->unordered_write = true;
1484    obj->unsync_access = true;
1485    obj->last_dt_idx = obj->dt_idx = UINT32_MAX; //TODO: unionize
1486 
1487    struct mem_alloc_info alloc_info = {
1488       .whandle = whandle,
1489       .need_dedicated = false,
1490       .export_types = ZINK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_BIT,
1491       .shared = templ->bind & PIPE_BIND_SHARED,
1492       .user_mem = user_mem
1493    };
1494 
1495    /* figure out aux plane count */
1496    if (whandle && whandle->plane >= util_format_get_num_planes(whandle->format))
1497       obj->is_aux = true;
1498    struct pipe_resource *pnext = templ->next;
1499    for (obj->plane_count = 1; pnext; obj->plane_count++, pnext = pnext->next) {
1500       struct zink_resource *next = zink_resource(pnext);
1501       if (!next->obj->is_aux)
1502          break;
1503    }
1504 
1505    if (!get_export_flags(screen, templ, &alloc_info)) {
1506       /* can't export anything, fail early */
1507       return NULL;
1508    }
1509 
1510    pipe_reference_init(&obj->reference, 1);
1511    if (loader_private) {
1512       obj->bo = CALLOC_STRUCT(zink_bo);
1513       if (!obj->bo) {
1514          mesa_loge("ZINK: failed to allocate obj->bo!");
1515          return NULL;
1516       }
1517 
1518       obj->transfer_dst = true;
1519       return obj;
1520    }
1521 
1522    enum resource_object_create_result create_result;
1523    if (templ->target == PIPE_BUFFER) {
1524       max_level = 1;
1525       create_result = create_buffer(screen, obj, templ, modifiers, modifiers_count, &alloc_info);
1526    } else {
1527       max_level = templ->last_level + 1;
1528       create_result = create_image(screen, obj, templ, linear, modifiers, modifiers_count,
1529                                    &alloc_info);
1530    }
1531 
1532    switch (create_result) {
1533    case roc_success:
1534       for (unsigned i = 0; i < max_level; i++)
1535          util_dynarray_init(&obj->copies[i], NULL);
1536       FALLTHROUGH;
1537    case roc_success_early_return:
1538       return obj;
1539 
1540    case roc_fail_and_cleanup_all:
1541       zink_bo_unref(screen, obj->bo);
1542       FALLTHROUGH;
1543    case roc_fail_and_cleanup_object:
1544       if (templ->target == PIPE_BUFFER) {
1545          VKSCR(DestroyBuffer)(screen->dev, obj->buffer, NULL);
1546          VKSCR(DestroyBuffer)(screen->dev, obj->storage_buffer, NULL);
1547       } else
1548          VKSCR(DestroyImage)(screen->dev, obj->image, NULL);
1549       FALLTHROUGH;
1550    case roc_fail_and_free_object:
1551       FREE(obj);
1552       return NULL;
1553    default:
1554       unreachable("Invalid create object result code");
1555    }
1556 }
1557 
1558 static struct pipe_resource *
resource_create(struct pipe_screen * pscreen,const struct pipe_resource * templ,struct winsys_handle * whandle,unsigned external_usage,const uint64_t * modifiers,int modifiers_count,const void * loader_private,const void * user_mem)1559 resource_create(struct pipe_screen *pscreen,
1560                 const struct pipe_resource *templ,
1561                 struct winsys_handle *whandle,
1562                 unsigned external_usage,
1563                 const uint64_t *modifiers, int modifiers_count,
1564                 const void *loader_private, const void *user_mem)
1565 {
1566    struct zink_screen *screen = zink_screen(pscreen);
1567    struct zink_resource *res = CALLOC_STRUCT_CL(zink_resource);
1568 
1569    if (!res) {
1570       mesa_loge("ZINK: failed to allocate res!");
1571       return NULL;
1572    }
1573 
1574    if (modifiers_count > 0 && screen->info.have_EXT_image_drm_format_modifier) {
1575       /* for rebinds */
1576       res->modifiers_count = modifiers_count;
1577       res->modifiers = mem_dup(modifiers, modifiers_count * sizeof(uint64_t));
1578       if (!res->modifiers) {
1579          FREE_CL(res);
1580          return NULL;
1581       }
1582    }
1583 
1584    res->base.b = *templ;
1585 
1586    bool allow_cpu_storage = (templ->target == PIPE_BUFFER) &&
1587                             (templ->usage != PIPE_USAGE_STREAM) &&
1588                             (templ->width0 < 0x1000);
1589    threaded_resource_init(&res->base.b, allow_cpu_storage);
1590    pipe_reference_init(&res->base.b.reference, 1);
1591    res->base.b.screen = pscreen;
1592 
1593    bool linear = false;
1594    struct pipe_resource templ2 = *templ;
1595    if (templ2.flags & PIPE_RESOURCE_FLAG_SPARSE &&
1596        (util_res_sample_count(templ) == 1 || screen->info.feats.features.shaderStorageImageMultisample))
1597       templ2.bind |= PIPE_BIND_SHADER_IMAGE;
1598    res->obj = resource_object_create(screen, &templ2, whandle, &linear, res->modifiers, res->modifiers_count, loader_private, user_mem);
1599    if (!res->obj) {
1600       free(res->modifiers);
1601       FREE_CL(res);
1602       return NULL;
1603    }
1604 
1605    res->queue = VK_QUEUE_FAMILY_IGNORED;
1606    res->internal_format = templ->format;
1607    if (templ->target == PIPE_BUFFER) {
1608       util_range_init(&res->valid_buffer_range);
1609       res->base.b.bind |= PIPE_BIND_SHADER_IMAGE;
1610       if (!screen->resizable_bar && templ->width0 >= 8196) {
1611          /* We don't want to evict buffers from VRAM by mapping them for CPU access,
1612           * because they might never be moved back again. If a buffer is large enough,
1613           * upload data by copying from a temporary GTT buffer. 8K might not seem much,
1614           * but there can be 100000 buffers.
1615           *
1616           * This tweak improves performance for viewperf.
1617           */
1618          res->base.b.flags |= PIPE_RESOURCE_FLAG_DONT_MAP_DIRECTLY;
1619       }
1620       if (zink_descriptor_mode == ZINK_DESCRIPTOR_MODE_DB)
1621          zink_resource_get_address(screen, res);
1622    } else {
1623       if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE)
1624          res->base.b.bind |= PIPE_BIND_SHADER_IMAGE;
1625       if (templ->flags & PIPE_RESOURCE_FLAG_SPARSE) {
1626          uint32_t count = 1;
1627          VKSCR(GetImageSparseMemoryRequirements)(screen->dev, res->obj->image, &count, &res->sparse);
1628          res->base.b.nr_sparse_levels = res->sparse.imageMipTailFirstLod;
1629       }
1630       res->format = zink_get_format(screen, templ->format);
1631       if (templ->target == PIPE_TEXTURE_1D || templ->target == PIPE_TEXTURE_1D_ARRAY) {
1632          res->need_2D = (screen->need_2D_zs && util_format_is_depth_or_stencil(templ->format)) ||
1633                         (screen->need_2D_sparse && (templ->flags & PIPE_RESOURCE_FLAG_SPARSE));
1634       }
1635       res->dmabuf = whandle && whandle->type == WINSYS_HANDLE_TYPE_FD;
1636       if (res->dmabuf)
1637          res->queue = VK_QUEUE_FAMILY_FOREIGN_EXT;
1638       res->layout = res->dmabuf ? VK_IMAGE_LAYOUT_PREINITIALIZED : VK_IMAGE_LAYOUT_UNDEFINED;
1639       res->linear = linear;
1640       res->aspect = aspect_from_format(templ->format);
1641    }
1642 
1643    if (loader_private) {
1644       if (templ->bind & PIPE_BIND_DISPLAY_TARGET) {
1645          /* backbuffer */
1646          res->obj->dt = zink_kopper_displaytarget_create(screen,
1647                                                          res->base.b.bind,
1648                                                          res->base.b.format,
1649                                                          templ->width0,
1650                                                          templ->height0,
1651                                                          64, loader_private,
1652                                                          &res->dt_stride);
1653          if (!res->obj->dt) {
1654             mesa_loge("zink: could not create swapchain");
1655             FREE(res->obj);
1656             free(res->modifiers);
1657             FREE_CL(res);
1658             return NULL;
1659          }
1660          struct kopper_displaytarget *cdt = res->obj->dt;
1661          if (cdt->swapchain->num_acquires) {
1662             /* this should be a reused swapchain after a MakeCurrent dance that deleted the original resource */
1663             for (unsigned i = 0; i < cdt->swapchain->num_images; i++) {
1664                if (!cdt->swapchain->images[i].acquired)
1665                   continue;
1666                res->obj->dt_idx = i;
1667                res->obj->image = cdt->swapchain->images[i].image;
1668                res->layout = cdt->swapchain->images[i].layout;
1669             }
1670          }
1671       } else {
1672          /* frontbuffer */
1673          struct zink_resource *back = (void*)loader_private;
1674          struct kopper_displaytarget *cdt = back->obj->dt;
1675          cdt->refcount++;
1676          assert(back->obj->dt);
1677          res->obj->dt = back->obj->dt;
1678       }
1679       struct kopper_displaytarget *cdt = res->obj->dt;
1680       if (zink_kopper_has_srgb(cdt))
1681          res->obj->vkflags |= VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT;
1682       if (cdt->swapchain->scci.flags == VK_SWAPCHAIN_CREATE_MUTABLE_FORMAT_BIT_KHR)
1683          res->obj->vkflags = VK_IMAGE_CREATE_MUTABLE_FORMAT_BIT | VK_IMAGE_CREATE_EXTENDED_USAGE_BIT;
1684       res->obj->vkusage = cdt->swapchain->scci.imageUsage;
1685       res->base.b.bind |= PIPE_BIND_DISPLAY_TARGET;
1686       res->linear = false;
1687       res->swapchain = true;
1688    }
1689 
1690    if (!res->obj->host_visible) {
1691       res->base.b.flags |= PIPE_RESOURCE_FLAG_DONT_MAP_DIRECTLY;
1692       res->base.allow_cpu_storage = false;
1693    }
1694    if (res->obj->is_buffer) {
1695       res->base.buffer_id_unique = util_idalloc_mt_alloc(&screen->buffer_ids);
1696       _mesa_hash_table_init(&res->bufferview_cache, NULL, NULL, equals_bvci);
1697       simple_mtx_init(&res->bufferview_mtx, mtx_plain);
1698    } else {
1699       _mesa_hash_table_init(&res->surface_cache, NULL, NULL, equals_ivci);
1700       simple_mtx_init(&res->surface_mtx, mtx_plain);
1701    }
1702    if (res->obj->exportable)
1703       res->base.b.bind |= ZINK_BIND_DMABUF;
1704    return &res->base.b;
1705 }
1706 
1707 static struct pipe_resource *
zink_resource_create(struct pipe_screen * pscreen,const struct pipe_resource * templ)1708 zink_resource_create(struct pipe_screen *pscreen,
1709                      const struct pipe_resource *templ)
1710 {
1711    return resource_create(pscreen, templ, NULL, 0, NULL, 0, NULL, NULL);
1712 }
1713 
1714 static struct pipe_resource *
zink_resource_create_with_modifiers(struct pipe_screen * pscreen,const struct pipe_resource * templ,const uint64_t * modifiers,int modifiers_count)1715 zink_resource_create_with_modifiers(struct pipe_screen *pscreen, const struct pipe_resource *templ,
1716                                     const uint64_t *modifiers, int modifiers_count)
1717 {
1718    return resource_create(pscreen, templ, NULL, 0, modifiers, modifiers_count, NULL, NULL);
1719 }
1720 
1721 static struct pipe_resource *
zink_resource_create_drawable(struct pipe_screen * pscreen,const struct pipe_resource * templ,const void * loader_private)1722 zink_resource_create_drawable(struct pipe_screen *pscreen,
1723                               const struct pipe_resource *templ,
1724                               const void *loader_private)
1725 {
1726    return resource_create(pscreen, templ, NULL, 0, NULL, 0, loader_private, NULL);
1727 }
1728 
1729 static bool
add_resource_bind(struct zink_context * ctx,struct zink_resource * res,unsigned bind)1730 add_resource_bind(struct zink_context *ctx, struct zink_resource *res, unsigned bind)
1731 {
1732    struct zink_screen *screen = zink_screen(ctx->base.screen);
1733    assert((res->base.b.bind & bind) == 0);
1734    res->base.b.bind |= bind;
1735    struct zink_resource_object *old_obj = res->obj;
1736    if (bind & ZINK_BIND_DMABUF && !res->modifiers_count && screen->info.have_EXT_image_drm_format_modifier) {
1737       res->modifiers_count = 1;
1738       res->modifiers = malloc(res->modifiers_count * sizeof(uint64_t));
1739       if (!res->modifiers) {
1740          mesa_loge("ZINK: failed to allocate res->modifiers!");
1741          return false;
1742       }
1743 
1744       res->modifiers[0] = DRM_FORMAT_MOD_LINEAR;
1745    }
1746    struct zink_resource_object *new_obj = resource_object_create(screen, &res->base.b, NULL, &res->linear, res->modifiers, res->modifiers_count, NULL, NULL);
1747    if (!new_obj) {
1748       debug_printf("new backing resource alloc failed!\n");
1749       res->base.b.bind &= ~bind;
1750       return false;
1751    }
1752    struct zink_resource staging = *res;
1753    staging.obj = old_obj;
1754    staging.all_binds = 0;
1755    res->layout = VK_IMAGE_LAYOUT_UNDEFINED;
1756    res->obj = new_obj;
1757    res->queue = VK_QUEUE_FAMILY_IGNORED;
1758    for (unsigned i = 0; i <= res->base.b.last_level; i++) {
1759       struct pipe_box box;
1760       u_box_3d(0, 0, 0,
1761                u_minify(res->base.b.width0, i),
1762                u_minify(res->base.b.height0, i), res->base.b.array_size, &box);
1763       box.depth = util_num_layers(&res->base.b, i);
1764       ctx->base.resource_copy_region(&ctx->base, &res->base.b, i, 0, 0, 0, &staging.base.b, i, &box);
1765    }
1766    if (old_obj->exportable) {
1767       simple_mtx_lock(&ctx->bs->exportable_lock);
1768       _mesa_set_remove_key(&ctx->bs->dmabuf_exports, &staging);
1769       simple_mtx_unlock(&ctx->bs->exportable_lock);
1770    }
1771    zink_resource_object_reference(screen, &old_obj, NULL);
1772    return true;
1773 }
1774 
1775 static bool
zink_resource_is_aux_plane(struct pipe_resource * pres)1776 zink_resource_is_aux_plane(struct pipe_resource *pres)
1777 {
1778    struct zink_resource *rsc = zink_resource(pres);
1779    return rsc->obj->is_aux;
1780 }
1781 
1782 static bool
zink_resource_get_param(struct pipe_screen * pscreen,struct pipe_context * pctx,struct pipe_resource * pres,unsigned plane,unsigned layer,unsigned level,enum pipe_resource_param param,unsigned handle_usage,uint64_t * value)1783 zink_resource_get_param(struct pipe_screen *pscreen, struct pipe_context *pctx,
1784                         struct pipe_resource *pres,
1785                         unsigned plane,
1786                         unsigned layer,
1787                         unsigned level,
1788                         enum pipe_resource_param param,
1789                         unsigned handle_usage,
1790                         uint64_t *value)
1791 {
1792    while (plane && pres->next && !zink_resource_is_aux_plane(pres->next)) {
1793       --plane;
1794       pres = pres->next;
1795    }
1796 
1797    struct zink_screen *screen = zink_screen(pscreen);
1798    struct zink_resource *res = zink_resource(pres);
1799    struct zink_resource_object *obj = res->obj;
1800    struct winsys_handle whandle;
1801    VkImageAspectFlags aspect;
1802    if (obj->modifier_aspect) {
1803       switch (plane) {
1804       case 0:
1805          aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_0_BIT_EXT;
1806          break;
1807       case 1:
1808          aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_1_BIT_EXT;
1809          break;
1810       case 2:
1811          aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_2_BIT_EXT;
1812          break;
1813       case 3:
1814          aspect = VK_IMAGE_ASPECT_MEMORY_PLANE_3_BIT_EXT;
1815          break;
1816       default:
1817          unreachable("how many planes you got in this thing?");
1818       }
1819    } else if (util_format_is_yuv(pres->format)) {
1820       aspect = VK_IMAGE_ASPECT_PLANE_0_BIT;
1821    } else {
1822       aspect = res->aspect;
1823    }
1824    switch (param) {
1825    case PIPE_RESOURCE_PARAM_NPLANES:
1826       if (screen->info.have_EXT_image_drm_format_modifier)
1827          *value = screen->base.get_dmabuf_modifier_planes(&screen->base, obj->modifier, res->internal_format);
1828       else
1829          *value = 1;
1830       break;
1831 
1832    case PIPE_RESOURCE_PARAM_STRIDE: {
1833       VkImageSubresource sub_res = {0};
1834       VkSubresourceLayout sub_res_layout = {0};
1835 
1836       sub_res.aspectMask = aspect;
1837 
1838       VKSCR(GetImageSubresourceLayout)(screen->dev, obj->image, &sub_res, &sub_res_layout);
1839 
1840       *value = sub_res_layout.rowPitch;
1841       break;
1842    }
1843 
1844    case PIPE_RESOURCE_PARAM_OFFSET: {
1845          VkImageSubresource isr = {
1846             aspect,
1847             level,
1848             layer
1849          };
1850          VkSubresourceLayout srl;
1851          VKSCR(GetImageSubresourceLayout)(screen->dev, obj->image, &isr, &srl);
1852          *value = srl.offset;
1853          break;
1854    }
1855 
1856    case PIPE_RESOURCE_PARAM_MODIFIER: {
1857       *value = obj->modifier;
1858       break;
1859    }
1860 
1861    case PIPE_RESOURCE_PARAM_LAYER_STRIDE: {
1862          VkImageSubresource isr = {
1863             aspect,
1864             level,
1865             layer
1866          };
1867          VkSubresourceLayout srl;
1868          VKSCR(GetImageSubresourceLayout)(screen->dev, obj->image, &isr, &srl);
1869          if (res->base.b.target == PIPE_TEXTURE_3D)
1870             *value = srl.depthPitch;
1871          else
1872             *value = srl.arrayPitch;
1873          break;
1874    }
1875 
1876       return false;
1877    case PIPE_RESOURCE_PARAM_HANDLE_TYPE_KMS:
1878    case PIPE_RESOURCE_PARAM_HANDLE_TYPE_SHARED:
1879    case PIPE_RESOURCE_PARAM_HANDLE_TYPE_FD: {
1880 #ifdef ZINK_USE_DMABUF
1881       memset(&whandle, 0, sizeof(whandle));
1882       if (param == PIPE_RESOURCE_PARAM_HANDLE_TYPE_SHARED)
1883          whandle.type = WINSYS_HANDLE_TYPE_SHARED;
1884       if (param == PIPE_RESOURCE_PARAM_HANDLE_TYPE_KMS)
1885          whandle.type = WINSYS_HANDLE_TYPE_KMS;
1886       else if (param == PIPE_RESOURCE_PARAM_HANDLE_TYPE_FD)
1887          whandle.type = WINSYS_HANDLE_TYPE_FD;
1888 
1889       if (!pscreen->resource_get_handle(pscreen, pctx, pres, &whandle, handle_usage))
1890          return false;
1891 
1892 #ifdef _WIN32
1893       *value = (uintptr_t)whandle.handle;
1894 #else
1895       *value = whandle.handle;
1896 #endif
1897       break;
1898 #else
1899       (void)whandle;
1900       return false;
1901 #endif
1902    }
1903    }
1904    return true;
1905 }
1906 
1907 static bool
zink_resource_get_handle(struct pipe_screen * pscreen,struct pipe_context * context,struct pipe_resource * tex,struct winsys_handle * whandle,unsigned usage)1908 zink_resource_get_handle(struct pipe_screen *pscreen,
1909                          struct pipe_context *context,
1910                          struct pipe_resource *tex,
1911                          struct winsys_handle *whandle,
1912                          unsigned usage)
1913 {
1914    if (tex->target == PIPE_BUFFER)
1915       tc_buffer_disable_cpu_storage(tex);
1916    if (whandle->type == WINSYS_HANDLE_TYPE_FD || whandle->type == WINSYS_HANDLE_TYPE_KMS) {
1917 #ifdef ZINK_USE_DMABUF
1918       while (whandle->plane && tex->next && !zink_resource_is_aux_plane(tex->next)) {
1919          tex = tex->next;
1920       }
1921 
1922       struct zink_resource *res = zink_resource(tex);
1923       struct zink_screen *screen = zink_screen(pscreen);
1924       struct zink_resource_object *obj = res->obj;
1925 
1926 #if !defined(_WIN32)
1927       if (whandle->type == WINSYS_HANDLE_TYPE_KMS && screen->drm_fd == -1) {
1928          whandle->handle = -1;
1929       } else {
1930          if (!res->obj->exportable) {
1931             assert(!zink_resource_usage_is_unflushed(res));
1932             if (!screen->info.have_EXT_image_drm_format_modifier) {
1933                static bool warned = false;
1934                warn_missing_feature(warned, "EXT_image_drm_format_modifier");
1935                return false;
1936             }
1937             unsigned bind = ZINK_BIND_DMABUF;
1938             if (!(res->base.b.bind & PIPE_BIND_SHARED))
1939                bind |= PIPE_BIND_SHARED;
1940             zink_screen_lock_context(screen);
1941             if (!add_resource_bind(screen->copy_context, res, bind)) {
1942                zink_screen_unlock_context(screen);
1943                return false;
1944             }
1945             if (res->all_binds)
1946                p_atomic_inc(&screen->image_rebind_counter);
1947             screen->copy_context->base.flush(&screen->copy_context->base, NULL, 0);
1948             zink_screen_unlock_context(screen);
1949             obj = res->obj;
1950          }
1951 
1952          VkMemoryGetFdInfoKHR fd_info = {0};
1953          int fd;
1954          fd_info.sType = VK_STRUCTURE_TYPE_MEMORY_GET_FD_INFO_KHR;
1955          fd_info.memory = zink_bo_get_mem(obj->bo);
1956          if (whandle->type == WINSYS_HANDLE_TYPE_FD)
1957             fd_info.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_DMA_BUF_BIT_EXT;
1958          else
1959             fd_info.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_FD_BIT;
1960          VkResult result = VKSCR(GetMemoryFdKHR)(screen->dev, &fd_info, &fd);
1961          if (result != VK_SUCCESS) {
1962             mesa_loge("ZINK: vkGetMemoryFdKHR failed");
1963             return false;
1964          }
1965          if (whandle->type == WINSYS_HANDLE_TYPE_KMS) {
1966             uint32_t h;
1967             bool ret = zink_bo_get_kms_handle(screen, obj->bo, fd, &h);
1968             close(fd);
1969             if (!ret)
1970                return false;
1971             fd = h;
1972          }
1973 
1974          whandle->handle = fd;
1975       }
1976 #else
1977       VkMemoryGetWin32HandleInfoKHR handle_info = {0};
1978       HANDLE handle;
1979       handle_info.sType = VK_STRUCTURE_TYPE_MEMORY_GET_WIN32_HANDLE_INFO_KHR;
1980       //TODO: remove for wsi
1981       handle_info.memory = zink_bo_get_mem(obj->bo);
1982       handle_info.handleType = VK_EXTERNAL_MEMORY_HANDLE_TYPE_OPAQUE_WIN32_BIT;
1983       VkResult result = VKSCR(GetMemoryWin32HandleKHR)(screen->dev, &handle_info, &handle);
1984       if (result != VK_SUCCESS)
1985          return false;
1986       whandle->handle = handle;
1987 #endif
1988       uint64_t value;
1989       zink_resource_get_param(pscreen, context, tex, 0, 0, 0, PIPE_RESOURCE_PARAM_MODIFIER, 0, &value);
1990       whandle->modifier = value;
1991       zink_resource_get_param(pscreen, context, tex, 0, 0, 0, PIPE_RESOURCE_PARAM_OFFSET, 0, &value);
1992       whandle->offset = value;
1993       zink_resource_get_param(pscreen, context, tex, 0, 0, 0, PIPE_RESOURCE_PARAM_STRIDE, 0, &value);
1994       whandle->stride = value;
1995 #else
1996       return false;
1997 #endif
1998    }
1999    return true;
2000 }
2001 
2002 static struct pipe_resource *
zink_resource_from_handle(struct pipe_screen * pscreen,const struct pipe_resource * templ,struct winsys_handle * whandle,unsigned usage)2003 zink_resource_from_handle(struct pipe_screen *pscreen,
2004                  const struct pipe_resource *templ,
2005                  struct winsys_handle *whandle,
2006                  unsigned usage)
2007 {
2008 #ifdef ZINK_USE_DMABUF
2009    if (whandle->modifier != DRM_FORMAT_MOD_INVALID &&
2010        !zink_screen(pscreen)->info.have_EXT_image_drm_format_modifier)
2011       return NULL;
2012 
2013    struct pipe_resource templ2 = *templ;
2014    if (templ->format == PIPE_FORMAT_NONE)
2015       templ2.format = whandle->format;
2016 
2017    uint64_t modifier = DRM_FORMAT_MOD_LINEAR;
2018    int modifier_count = 1;
2019    if (whandle->modifier != DRM_FORMAT_MOD_INVALID)
2020       modifier = whandle->modifier;
2021    else {
2022       if (!zink_screen(pscreen)->driver_workarounds.can_do_invalid_linear_modifier) {
2023          mesa_loge("zink: display server doesn't support DRI3 modifiers and driver can't handle INVALID<->LINEAR!");
2024          return NULL;
2025       }
2026       whandle->modifier = modifier;
2027    }
2028    templ2.bind |= ZINK_BIND_DMABUF;
2029    struct pipe_resource *pres = resource_create(pscreen, &templ2, whandle, usage, &modifier, modifier_count, NULL, NULL);
2030    if (pres) {
2031       struct zink_resource *res = zink_resource(pres);
2032       if (pres->target != PIPE_BUFFER)
2033          res->valid = true;
2034       else
2035          tc_buffer_disable_cpu_storage(pres);
2036       res->internal_format = whandle->format;
2037    }
2038    return pres;
2039 #else
2040    return NULL;
2041 #endif
2042 }
2043 
2044 static struct pipe_resource *
zink_resource_from_user_memory(struct pipe_screen * pscreen,const struct pipe_resource * templ,void * user_memory)2045 zink_resource_from_user_memory(struct pipe_screen *pscreen,
2046                  const struct pipe_resource *templ,
2047                  void *user_memory)
2048 {
2049    struct zink_screen *screen = zink_screen(pscreen);
2050    VkDeviceSize alignMask = screen->info.ext_host_mem_props.minImportedHostPointerAlignment - 1;
2051 
2052    /* Validate the user_memory pointer and fail early.
2053     * minImportedHostPointerAlignment is required to be POT */
2054    if (((uintptr_t)user_memory) & alignMask)
2055       return NULL;
2056 
2057    return resource_create(pscreen, templ, NULL, 0, NULL, 0, NULL, user_memory);
2058 }
2059 
2060 struct zink_memory_object {
2061    struct pipe_memory_object b;
2062    struct winsys_handle whandle;
2063 };
2064 
2065 static struct pipe_memory_object *
zink_memobj_create_from_handle(struct pipe_screen * pscreen,struct winsys_handle * whandle,bool dedicated)2066 zink_memobj_create_from_handle(struct pipe_screen *pscreen, struct winsys_handle *whandle, bool dedicated)
2067 {
2068    struct zink_memory_object *memobj = CALLOC_STRUCT(zink_memory_object);
2069    if (!memobj)
2070       return NULL;
2071    memcpy(&memobj->whandle, whandle, sizeof(struct winsys_handle));
2072    memobj->whandle.type = ZINK_EXTERNAL_MEMORY_HANDLE;
2073 
2074 #ifdef ZINK_USE_DMABUF
2075 
2076 #if !defined(_WIN32)
2077    memobj->whandle.handle = os_dupfd_cloexec(whandle->handle);
2078 #else
2079    HANDLE source_target = GetCurrentProcess();
2080    HANDLE out_handle;
2081 
2082    DuplicateHandle(source_target, whandle->handle, source_target, &out_handle, 0, false, DUPLICATE_SAME_ACCESS);
2083    memobj->whandle.handle = out_handle;
2084 
2085 #endif /* _WIN32 */
2086 #endif /* ZINK_USE_DMABUF */
2087 
2088    return (struct pipe_memory_object *)memobj;
2089 }
2090 
2091 static void
zink_memobj_destroy(struct pipe_screen * pscreen,struct pipe_memory_object * pmemobj)2092 zink_memobj_destroy(struct pipe_screen *pscreen, struct pipe_memory_object *pmemobj)
2093 {
2094 #ifdef ZINK_USE_DMABUF
2095    struct zink_memory_object *memobj = (struct zink_memory_object *)pmemobj;
2096 
2097 #if !defined(_WIN32)
2098    close(memobj->whandle.handle);
2099 #else
2100    CloseHandle(memobj->whandle.handle);
2101 #endif /* _WIN32 */
2102 #endif /* ZINK_USE_DMABUF */
2103 
2104    FREE(pmemobj);
2105 }
2106 
2107 static struct pipe_resource *
zink_resource_from_memobj(struct pipe_screen * pscreen,const struct pipe_resource * templ,struct pipe_memory_object * pmemobj,uint64_t offset)2108 zink_resource_from_memobj(struct pipe_screen *pscreen,
2109                           const struct pipe_resource *templ,
2110                           struct pipe_memory_object *pmemobj,
2111                           uint64_t offset)
2112 {
2113    struct zink_memory_object *memobj = (struct zink_memory_object *)pmemobj;
2114 
2115    struct pipe_resource *pres = resource_create(pscreen, templ, &memobj->whandle, 0, NULL, 0, NULL, NULL);
2116    if (pres) {
2117       if (pres->target != PIPE_BUFFER)
2118          zink_resource(pres)->valid = true;
2119       else
2120          tc_buffer_disable_cpu_storage(pres);
2121    }
2122    return pres;
2123 }
2124 
2125 static bool
invalidate_buffer(struct zink_context * ctx,struct zink_resource * res)2126 invalidate_buffer(struct zink_context *ctx, struct zink_resource *res)
2127 {
2128    struct zink_screen *screen = zink_screen(ctx->base.screen);
2129 
2130    assert(res->base.b.target == PIPE_BUFFER);
2131 
2132    if (res->base.b.flags & PIPE_RESOURCE_FLAG_SPARSE)
2133       return false;
2134 
2135    struct pipe_box box;
2136    u_box_3d(0, 0, 0, res->base.b.width0, 0, 0, &box);
2137    if (res->valid_buffer_range.start > res->valid_buffer_range.end &&
2138        !zink_resource_copy_box_intersects(res, 0, &box))
2139       return false;
2140 
2141    if (res->so_valid)
2142       ctx->dirty_so_targets = true;
2143    /* force counter buffer reset */
2144    res->so_valid = false;
2145 
2146    util_range_set_empty(&res->valid_buffer_range);
2147    if (!zink_resource_has_usage(res))
2148       return false;
2149 
2150    struct zink_resource_object *new_obj = resource_object_create(screen, &res->base.b, NULL, NULL, NULL, 0, NULL, 0);
2151    if (!new_obj) {
2152       debug_printf("new backing resource alloc failed!\n");
2153       return false;
2154    }
2155    bool needs_bda = !!res->obj->bda;
2156    /* this ref must be transferred before rebind or else BOOM */
2157    zink_batch_reference_resource_move(ctx, res);
2158    res->obj = new_obj;
2159    res->queue = VK_QUEUE_FAMILY_IGNORED;
2160    if (needs_bda)
2161       zink_resource_get_address(screen, res);
2162    zink_resource_rebind(ctx, res);
2163    return true;
2164 }
2165 
2166 
2167 static void
zink_resource_invalidate(struct pipe_context * pctx,struct pipe_resource * pres)2168 zink_resource_invalidate(struct pipe_context *pctx, struct pipe_resource *pres)
2169 {
2170    if (pres->target == PIPE_BUFFER)
2171       invalidate_buffer(zink_context(pctx), zink_resource(pres));
2172    else {
2173       struct zink_resource *res = zink_resource(pres);
2174       if (res->valid && res->fb_bind_count)
2175          zink_context(pctx)->rp_loadop_changed = true;
2176       res->valid = false;
2177    }
2178 }
2179 
2180 static void
zink_transfer_copy_bufimage(struct zink_context * ctx,struct zink_resource * dst,struct zink_resource * src,struct zink_transfer * trans)2181 zink_transfer_copy_bufimage(struct zink_context *ctx,
2182                             struct zink_resource *dst,
2183                             struct zink_resource *src,
2184                             struct zink_transfer *trans)
2185 {
2186    assert((trans->base.b.usage & (PIPE_MAP_DEPTH_ONLY | PIPE_MAP_STENCIL_ONLY)) !=
2187           (PIPE_MAP_DEPTH_ONLY | PIPE_MAP_STENCIL_ONLY));
2188 
2189    bool buf2img = src->base.b.target == PIPE_BUFFER;
2190 
2191    struct pipe_box box = trans->base.b.box;
2192    int x = box.x;
2193    if (buf2img)
2194       box.x = trans->offset;
2195 
2196    assert(dst->obj->transfer_dst);
2197    zink_copy_image_buffer(ctx, dst, src, trans->base.b.level, buf2img ? x : 0,
2198                            box.y, box.z, trans->base.b.level, &box, trans->base.b.usage);
2199 }
2200 
2201 ALWAYS_INLINE static void
align_offset_size(const VkDeviceSize alignment,VkDeviceSize * offset,VkDeviceSize * size,VkDeviceSize obj_size)2202 align_offset_size(const VkDeviceSize alignment, VkDeviceSize *offset, VkDeviceSize *size, VkDeviceSize obj_size)
2203 {
2204    VkDeviceSize align = *offset % alignment;
2205    if (alignment - 1 > *offset)
2206       *offset = 0;
2207    else
2208       *offset -= align, *size += align;
2209    align = alignment - (*size % alignment);
2210    if (*offset + *size + align > obj_size)
2211       *size = obj_size - *offset;
2212    else
2213       *size += align;
2214 }
2215 
2216 VkMappedMemoryRange
zink_resource_init_mem_range(struct zink_screen * screen,struct zink_resource_object * obj,VkDeviceSize offset,VkDeviceSize size)2217 zink_resource_init_mem_range(struct zink_screen *screen, struct zink_resource_object *obj, VkDeviceSize offset, VkDeviceSize size)
2218 {
2219    assert(obj->size);
2220    align_offset_size(screen->info.props.limits.nonCoherentAtomSize, &offset, &size, obj->size);
2221    VkMappedMemoryRange range = {
2222       VK_STRUCTURE_TYPE_MAPPED_MEMORY_RANGE,
2223       NULL,
2224       zink_bo_get_mem(obj->bo),
2225       offset,
2226       size
2227    };
2228    assert(range.size);
2229    return range;
2230 }
2231 
2232 static void *
map_resource(struct zink_screen * screen,struct zink_resource * res)2233 map_resource(struct zink_screen *screen, struct zink_resource *res)
2234 {
2235    assert(res->obj->host_visible);
2236    return zink_bo_map(screen, res->obj->bo);
2237 }
2238 
2239 static void
unmap_resource(struct zink_screen * screen,struct zink_resource * res)2240 unmap_resource(struct zink_screen *screen, struct zink_resource *res)
2241 {
2242    zink_bo_unmap(screen, res->obj->bo);
2243 }
2244 
2245 static struct zink_transfer *
create_transfer(struct zink_context * ctx,struct pipe_resource * pres,unsigned usage,const struct pipe_box * box)2246 create_transfer(struct zink_context *ctx, struct pipe_resource *pres, unsigned usage, const struct pipe_box *box)
2247 {
2248    struct zink_transfer *trans;
2249 
2250    if (usage & PIPE_MAP_THREAD_SAFE)
2251       trans = calloc(1, sizeof(*trans));
2252    else if (usage & TC_TRANSFER_MAP_THREADED_UNSYNC)
2253       trans = slab_zalloc(&ctx->transfer_pool_unsync);
2254    else
2255       trans = slab_zalloc(&ctx->transfer_pool);
2256    if (!trans)
2257       return NULL;
2258 
2259    pipe_resource_reference(&trans->base.b.resource, pres);
2260 
2261    trans->base.b.usage = usage;
2262    trans->base.b.box = *box;
2263    return trans;
2264 }
2265 
2266 static void
destroy_transfer(struct zink_context * ctx,struct zink_transfer * trans)2267 destroy_transfer(struct zink_context *ctx, struct zink_transfer *trans)
2268 {
2269    if (trans->base.b.usage & PIPE_MAP_THREAD_SAFE) {
2270       free(trans);
2271    } else {
2272       /* Don't use pool_transfers_unsync. We are always in the driver
2273        * thread. Freeing an object into a different pool is allowed.
2274        */
2275       slab_free(&ctx->transfer_pool, trans);
2276    }
2277 }
2278 
2279 static void *
zink_buffer_map(struct pipe_context * pctx,struct pipe_resource * pres,unsigned level,unsigned usage,const struct pipe_box * box,struct pipe_transfer ** transfer)2280 zink_buffer_map(struct pipe_context *pctx,
2281                     struct pipe_resource *pres,
2282                     unsigned level,
2283                     unsigned usage,
2284                     const struct pipe_box *box,
2285                     struct pipe_transfer **transfer)
2286 {
2287    struct zink_context *ctx = zink_context(pctx);
2288    struct zink_screen *screen = zink_screen(pctx->screen);
2289    struct zink_resource *res = zink_resource(pres);
2290    struct zink_transfer *trans = create_transfer(ctx, pres, usage, box);
2291    if (!trans)
2292       return NULL;
2293 
2294    void *ptr = NULL;
2295 
2296    if (res->base.is_user_ptr)
2297       usage |= PIPE_MAP_PERSISTENT;
2298 
2299    /* See if the buffer range being mapped has never been initialized,
2300     * in which case it can be mapped unsynchronized. */
2301    if (!(usage & (PIPE_MAP_UNSYNCHRONIZED | TC_TRANSFER_MAP_NO_INFER_UNSYNCHRONIZED)) &&
2302        usage & PIPE_MAP_WRITE && !res->base.is_shared &&
2303        !util_ranges_intersect(&res->valid_buffer_range, box->x, box->x + box->width) &&
2304        !zink_resource_copy_box_intersects(res, 0, box)) {
2305       usage |= PIPE_MAP_UNSYNCHRONIZED;
2306    }
2307 
2308    /* If discarding the entire range, discard the whole resource instead. */
2309    if (usage & PIPE_MAP_DISCARD_RANGE && box->x == 0 && box->width == res->base.b.width0) {
2310       usage |= PIPE_MAP_DISCARD_WHOLE_RESOURCE;
2311    }
2312 
2313    /* If a buffer in VRAM is too large and the range is discarded, don't
2314     * map it directly. This makes sure that the buffer stays in VRAM.
2315     */
2316    bool force_discard_range = false;
2317    if (usage & (PIPE_MAP_DISCARD_WHOLE_RESOURCE | PIPE_MAP_DISCARD_RANGE) &&
2318        !(usage & PIPE_MAP_PERSISTENT) &&
2319        res->base.b.flags & PIPE_RESOURCE_FLAG_DONT_MAP_DIRECTLY) {
2320       usage &= ~(PIPE_MAP_DISCARD_WHOLE_RESOURCE | PIPE_MAP_UNSYNCHRONIZED);
2321       usage |= PIPE_MAP_DISCARD_RANGE;
2322       force_discard_range = true;
2323    }
2324 
2325    if (usage & PIPE_MAP_DISCARD_WHOLE_RESOURCE &&
2326        !(usage & (PIPE_MAP_UNSYNCHRONIZED | TC_TRANSFER_MAP_NO_INVALIDATE))) {
2327       assert(usage & PIPE_MAP_WRITE);
2328 
2329       if (invalidate_buffer(ctx, res)) {
2330          /* At this point, the buffer is always idle. */
2331          usage |= PIPE_MAP_UNSYNCHRONIZED;
2332       } else {
2333          /* Fall back to a temporary buffer. */
2334          usage |= PIPE_MAP_DISCARD_RANGE;
2335       }
2336    }
2337 
2338    unsigned map_offset = box->x;
2339    if (usage & PIPE_MAP_DISCARD_RANGE &&
2340         (!res->obj->host_visible ||
2341         !(usage & (PIPE_MAP_UNSYNCHRONIZED | PIPE_MAP_PERSISTENT)))) {
2342 
2343       /* Check if mapping this buffer would cause waiting for the GPU.
2344        */
2345 
2346       if (!res->obj->host_visible || force_discard_range ||
2347           !zink_resource_usage_check_completion(screen, res, ZINK_RESOURCE_ACCESS_RW)) {
2348          /* Do a wait-free write-only transfer using a temporary buffer. */
2349          unsigned offset;
2350 
2351          /* If we are not called from the driver thread, we have
2352           * to use the uploader from u_threaded_context, which is
2353           * local to the calling thread.
2354           */
2355          struct u_upload_mgr *mgr;
2356          if (usage & TC_TRANSFER_MAP_THREADED_UNSYNC)
2357             mgr = ctx->tc->base.stream_uploader;
2358          else
2359             mgr = ctx->base.stream_uploader;
2360          u_upload_alloc(mgr, 0, box->width,
2361                      screen->info.props.limits.minMemoryMapAlignment, &offset,
2362                      (struct pipe_resource **)&trans->staging_res, (void **)&ptr);
2363          res = zink_resource(trans->staging_res);
2364          trans->offset = offset;
2365          usage |= PIPE_MAP_UNSYNCHRONIZED;
2366          ptr = ((uint8_t *)ptr);
2367       } else {
2368          /* At this point, the buffer is always idle (we checked it above). */
2369          usage |= PIPE_MAP_UNSYNCHRONIZED;
2370       }
2371    } else if (usage & ZINK_MAP_QBO) {
2372       usage |= PIPE_MAP_UNSYNCHRONIZED;
2373    } else if (usage & PIPE_MAP_DONTBLOCK) {
2374       /* sparse/device-local will always need to wait since it has to copy */
2375       if (!res->obj->host_visible)
2376          goto success;
2377       if (!zink_resource_usage_check_completion(screen, res, ZINK_RESOURCE_ACCESS_WRITE))
2378          goto success;
2379       usage |= PIPE_MAP_UNSYNCHRONIZED;
2380    } else if (((usage & PIPE_MAP_READ) && !(usage & PIPE_MAP_PERSISTENT) &&
2381                ((screen->info.mem_props.memoryTypes[res->obj->bo->base.base.placement].propertyFlags & VK_STAGING_RAM) != VK_STAGING_RAM)) ||
2382               !res->obj->host_visible) {
2383       /* any read, non-HV write, or unmappable that reaches this point needs staging */
2384       if ((usage & PIPE_MAP_READ) || !res->obj->host_visible || res->base.b.flags & PIPE_RESOURCE_FLAG_DONT_MAP_DIRECTLY) {
2385 overwrite:
2386          trans->offset = box->x % MAX2(screen->info.props.limits.minMemoryMapAlignment, 1 << MIN_SLAB_ORDER);
2387          trans->staging_res = pipe_buffer_create(&screen->base, PIPE_BIND_LINEAR, PIPE_USAGE_STAGING, box->width + trans->offset);
2388          if (!trans->staging_res)
2389             goto fail;
2390          struct zink_resource *staging_res = zink_resource(trans->staging_res);
2391          if (usage & (PIPE_MAP_THREAD_SAFE | PIPE_MAP_UNSYNCHRONIZED | TC_TRANSFER_MAP_THREADED_UNSYNC)) {
2392             assert(ctx != screen->copy_context);
2393             /* this map can't access the passed context: use the copy context */
2394             zink_screen_lock_context(screen);
2395             ctx = screen->copy_context;
2396          }
2397          if (usage & PIPE_MAP_READ)
2398             zink_copy_buffer(ctx, staging_res, res, trans->offset, box->x, box->width);
2399          res = staging_res;
2400          usage &= ~PIPE_MAP_UNSYNCHRONIZED;
2401          map_offset = trans->offset;
2402       }
2403    }
2404 
2405    if (!(usage & PIPE_MAP_UNSYNCHRONIZED)) {
2406       if (usage & PIPE_MAP_WRITE) {
2407          if (!(usage & PIPE_MAP_READ)) {
2408             zink_resource_usage_try_wait(ctx, res, ZINK_RESOURCE_ACCESS_RW);
2409             if (zink_resource_has_unflushed_usage(res))
2410                goto overwrite;
2411          }
2412          zink_resource_usage_wait(ctx, res, ZINK_RESOURCE_ACCESS_RW);
2413       } else
2414          zink_resource_usage_wait(ctx, res, ZINK_RESOURCE_ACCESS_WRITE);
2415       if (!res->real_buffer_range) {
2416          res->obj->access = 0;
2417          res->obj->access_stage = 0;
2418          res->obj->last_write = 0;
2419          zink_resource_copies_reset(res);
2420       }
2421    }
2422 
2423    if (!ptr) {
2424       /* if writing to a streamout buffer, ensure synchronization next time it's used */
2425       if (usage & PIPE_MAP_WRITE && res->so_valid) {
2426          ctx->dirty_so_targets = true;
2427          /* force counter buffer reset */
2428          res->so_valid = false;
2429       }
2430       ptr = map_resource(screen, res);
2431       if (!ptr)
2432          goto fail;
2433       ptr = ((uint8_t *)ptr) + map_offset;
2434    }
2435 
2436    if (!res->obj->coherent
2437 #if defined(MVK_VERSION)
2438       // Work around for MoltenVk limitation specifically on coherent memory
2439       // MoltenVk returns blank memory ranges when there should be data present
2440       // This is a known limitation of MoltenVK.
2441       // See https://github.com/KhronosGroup/MoltenVK/blob/master/Docs/MoltenVK_Runtime_UserGuide.md#known-moltenvk-limitations
2442 
2443        || screen->instance_info.have_MVK_moltenvk
2444 #endif
2445       ) {
2446       VkDeviceSize size = box->width;
2447       VkDeviceSize offset = res->obj->offset + trans->offset;
2448       VkMappedMemoryRange range = zink_resource_init_mem_range(screen, res->obj, offset, size);
2449       if (VKSCR(InvalidateMappedMemoryRanges)(screen->dev, 1, &range) != VK_SUCCESS) {
2450          mesa_loge("ZINK: vkInvalidateMappedMemoryRanges failed");
2451          zink_bo_unmap(screen, res->obj->bo);
2452          goto fail;
2453       }
2454    }
2455    trans->base.b.usage = usage;
2456    if (usage & PIPE_MAP_WRITE) {
2457       util_range_add(&res->base.b, &res->valid_buffer_range, box->x, box->x + box->width);
2458 
2459       struct zink_resource *orig_res = zink_resource(trans->base.b.resource);
2460       util_range_add(&orig_res->base.b, &orig_res->valid_buffer_range, box->x, box->x + box->width);
2461       if (orig_res->real_buffer_range)
2462          util_range_add(&orig_res->base.b, orig_res->real_buffer_range, box->x, box->x + box->width);
2463    }
2464 
2465 success:
2466    /* ensure the copy context gets unlocked */
2467    if (ctx == screen->copy_context)
2468       zink_screen_unlock_context(screen);
2469    *transfer = &trans->base.b;
2470    return ptr;
2471 
2472 fail:
2473    if (ctx == screen->copy_context)
2474       zink_screen_unlock_context(screen);
2475    destroy_transfer(ctx, trans);
2476    return NULL;
2477 }
2478 
2479 static void *
zink_image_map(struct pipe_context * pctx,struct pipe_resource * pres,unsigned level,unsigned usage,const struct pipe_box * box,struct pipe_transfer ** transfer)2480 zink_image_map(struct pipe_context *pctx,
2481                   struct pipe_resource *pres,
2482                   unsigned level,
2483                   unsigned usage,
2484                   const struct pipe_box *box,
2485                   struct pipe_transfer **transfer)
2486 {
2487    struct zink_context *ctx = zink_context(pctx);
2488    struct zink_screen *screen = zink_screen(pctx->screen);
2489    struct zink_resource *res = zink_resource(pres);
2490    struct zink_transfer *trans = create_transfer(ctx, pres, usage, box);
2491    if (!trans)
2492       return NULL;
2493 
2494    trans->base.b.level = level;
2495    if (zink_is_swapchain(res))
2496       /* this is probably a multi-chain which has already been acquired */
2497       zink_kopper_acquire(ctx, res, 0);
2498 
2499    void *ptr;
2500    if (!(usage & PIPE_MAP_UNSYNCHRONIZED)) {
2501       if (usage & PIPE_MAP_WRITE && !(usage & PIPE_MAP_READ))
2502          /* this is like a blit, so we can potentially dump some clears or maybe we have to  */
2503          zink_fb_clears_apply_or_discard(ctx, pres, zink_rect_from_box(box), false);
2504       else if (usage & PIPE_MAP_READ)
2505          /* if the map region intersects with any clears then we have to apply them */
2506          zink_fb_clears_apply_region(ctx, pres, zink_rect_from_box(box));
2507    }
2508    if (!res->linear || !res->obj->host_visible) {
2509       enum pipe_format format = pres->format;
2510       if (usage & PIPE_MAP_DEPTH_ONLY)
2511          format = util_format_get_depth_only(pres->format);
2512       else if (usage & PIPE_MAP_STENCIL_ONLY)
2513          format = PIPE_FORMAT_S8_UINT;
2514       trans->base.b.stride = util_format_get_stride(format, box->width);
2515       trans->base.b.layer_stride = util_format_get_2d_size(format,
2516                                                          trans->base.b.stride,
2517                                                          box->height);
2518 
2519       struct pipe_resource templ = *pres;
2520       templ.next = NULL;
2521       templ.format = format;
2522       templ.usage = usage & PIPE_MAP_READ ? PIPE_USAGE_STAGING : PIPE_USAGE_STREAM;
2523       templ.target = PIPE_BUFFER;
2524       templ.bind = PIPE_BIND_LINEAR;
2525       templ.width0 = trans->base.b.layer_stride * box->depth;
2526       templ.height0 = templ.depth0 = 0;
2527       templ.last_level = 0;
2528       templ.array_size = 1;
2529       templ.flags = 0;
2530 
2531       trans->staging_res = zink_resource_create(pctx->screen, &templ);
2532       if (!trans->staging_res)
2533          goto fail;
2534 
2535       struct zink_resource *staging_res = zink_resource(trans->staging_res);
2536 
2537       if (usage & PIPE_MAP_READ) {
2538          assert(!(usage & TC_TRANSFER_MAP_THREADED_UNSYNC));
2539          /* force multi-context sync */
2540          if (zink_resource_usage_is_unflushed_write(res))
2541             zink_resource_usage_wait(ctx, res, ZINK_RESOURCE_ACCESS_WRITE);
2542          zink_transfer_copy_bufimage(ctx, staging_res, res, trans);
2543          /* need to wait for rendering to finish */
2544          zink_fence_wait(pctx);
2545       }
2546 
2547       ptr = map_resource(screen, staging_res);
2548    } else {
2549       assert(res->linear);
2550       ptr = map_resource(screen, res);
2551       if (!ptr)
2552          goto fail;
2553       if (zink_resource_has_usage(res)) {
2554          assert(!(usage & PIPE_MAP_UNSYNCHRONIZED));
2555          if (usage & PIPE_MAP_WRITE)
2556             zink_fence_wait(pctx);
2557          else
2558             zink_resource_usage_wait(ctx, res, ZINK_RESOURCE_ACCESS_WRITE);
2559       }
2560       VkImageSubresource isr = {
2561          res->modifiers ? res->obj->modifier_aspect : res->aspect,
2562          level,
2563          0
2564       };
2565       VkSubresourceLayout srl;
2566       VKSCR(GetImageSubresourceLayout)(screen->dev, res->obj->image, &isr, &srl);
2567       trans->base.b.stride = srl.rowPitch;
2568       if (res->base.b.target == PIPE_TEXTURE_3D)
2569          trans->base.b.layer_stride = srl.depthPitch;
2570       else
2571          trans->base.b.layer_stride = srl.arrayPitch;
2572       trans->offset = srl.offset;
2573       trans->depthPitch = srl.depthPitch;
2574       const struct util_format_description *desc = util_format_description(res->base.b.format);
2575       unsigned offset = srl.offset +
2576                         box->z * srl.depthPitch +
2577                         (box->y / desc->block.height) * srl.rowPitch +
2578                         (box->x / desc->block.width) * (desc->block.bits / 8);
2579       if (!res->obj->coherent) {
2580          VkDeviceSize size = (VkDeviceSize)box->width * box->height * desc->block.bits / 8;
2581          VkMappedMemoryRange range = zink_resource_init_mem_range(screen, res->obj, res->obj->offset + offset, size);
2582          if (VKSCR(FlushMappedMemoryRanges)(screen->dev, 1, &range) != VK_SUCCESS) {
2583             mesa_loge("ZINK: vkFlushMappedMemoryRanges failed");
2584          }
2585       }
2586       ptr = ((uint8_t *)ptr) + offset;
2587    }
2588    if (!ptr)
2589       goto fail;
2590    if (usage & PIPE_MAP_WRITE) {
2591       if (!res->valid && res->fb_bind_count) {
2592          assert(!(usage & PIPE_MAP_UNSYNCHRONIZED));
2593          ctx->rp_loadop_changed = true;
2594       }
2595       res->valid = true;
2596    }
2597 
2598    if (sizeof(void*) == 4)
2599       trans->base.b.usage |= ZINK_MAP_TEMPORARY;
2600 
2601    *transfer = &trans->base.b;
2602    return ptr;
2603 
2604 fail:
2605    destroy_transfer(ctx, trans);
2606    return NULL;
2607 }
2608 
2609 static void
zink_image_subdata(struct pipe_context * pctx,struct pipe_resource * pres,unsigned level,unsigned usage,const struct pipe_box * box,const void * data,unsigned stride,uintptr_t layer_stride)2610 zink_image_subdata(struct pipe_context *pctx,
2611                   struct pipe_resource *pres,
2612                   unsigned level,
2613                   unsigned usage,
2614                   const struct pipe_box *box,
2615                   const void *data,
2616                   unsigned stride,
2617                   uintptr_t layer_stride)
2618 {
2619    struct zink_screen *screen = zink_screen(pctx->screen);
2620    struct zink_context *ctx = zink_context(pctx);
2621    struct zink_resource *res = zink_resource(pres);
2622 
2623    /* flush clears to avoid subdata conflict */
2624    if (!(usage & TC_TRANSFER_MAP_THREADED_UNSYNC) &&
2625        (res->obj->vkusage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT))
2626       zink_fb_clears_apply_or_discard(ctx, pres, zink_rect_from_box(box), false);
2627    /* only use HIC if supported on image and no pending usage */
2628    while (res->obj->vkusage & VK_IMAGE_USAGE_HOST_TRANSFER_BIT_EXT &&
2629           zink_resource_usage_check_completion(screen, res, ZINK_RESOURCE_ACCESS_RW)) {
2630       /* uninit images are always supported */
2631       bool change_layout = res->layout == VK_IMAGE_LAYOUT_UNDEFINED || res->layout == VK_IMAGE_LAYOUT_PREINITIALIZED;
2632       if (!change_layout) {
2633          /* image in some other layout: test for support */
2634          bool can_copy_layout = false;
2635          for (unsigned i = 0; i < screen->info.hic_props.copyDstLayoutCount; i++) {
2636             if (screen->info.hic_props.pCopyDstLayouts[i] == res->layout) {
2637                can_copy_layout = true;
2638                break;
2639             }
2640          }
2641          /* some layouts don't permit HIC copies */
2642          if (!can_copy_layout)
2643             break;
2644       }
2645       bool is_arrayed = false;
2646       switch (pres->target) {
2647       case PIPE_TEXTURE_1D_ARRAY:
2648       case PIPE_TEXTURE_2D_ARRAY:
2649       case PIPE_TEXTURE_CUBE:
2650       case PIPE_TEXTURE_CUBE_ARRAY:
2651          is_arrayed = true;
2652          break;
2653       default: break;
2654       }
2655       /* recalc strides into texel strides because HIC spec is insane */
2656       unsigned vk_stride = util_format_get_stride(pres->format, 1);
2657       stride /= vk_stride;
2658       unsigned vk_layer_stride = util_format_get_2d_size(pres->format, stride, 1) * vk_stride;
2659       layer_stride /= vk_layer_stride;
2660 
2661       VkHostImageLayoutTransitionInfoEXT t = {
2662          VK_STRUCTURE_TYPE_HOST_IMAGE_LAYOUT_TRANSITION_INFO_EXT,
2663          NULL,
2664          res->obj->image,
2665          res->layout,
2666          /* GENERAL support is guaranteed */
2667          VK_IMAGE_LAYOUT_GENERAL,
2668          {res->aspect, 0, VK_REMAINING_MIP_LEVELS, 0, VK_REMAINING_ARRAY_LAYERS}
2669       };
2670       /* only pre-transition uninit images to avoid thrashing */
2671       if (change_layout) {
2672          VKSCR(TransitionImageLayoutEXT)(screen->dev, 1, &t);
2673          res->layout = VK_IMAGE_LAYOUT_GENERAL;
2674       }
2675       VkMemoryToImageCopyEXT region = {
2676          VK_STRUCTURE_TYPE_MEMORY_TO_IMAGE_COPY_EXT,
2677          NULL,
2678          data,
2679          stride,
2680          layer_stride,
2681          {res->aspect, level, is_arrayed ? box->z : 0, is_arrayed ? box->depth : 1},
2682          {box->x, box->y, is_arrayed ? 0 : box->z},
2683          {box->width, box->height, is_arrayed ? 1 : box->depth}
2684       };
2685       VkCopyMemoryToImageInfoEXT copy = {
2686          VK_STRUCTURE_TYPE_COPY_MEMORY_TO_IMAGE_INFO_EXT,
2687          NULL,
2688          0,
2689          res->obj->image,
2690          res->layout,
2691          1,
2692          &region
2693       };
2694       VKSCR(CopyMemoryToImageEXT)(screen->dev, &copy);
2695       if (change_layout && screen->can_hic_shader_read && !pres->last_level && !box->x && !box->y && !box->z &&
2696           box->width == pres->width0 && box->height == pres->height0 &&
2697           ((is_arrayed && box->depth == pres->array_size) || (!is_arrayed && box->depth == pres->depth0))) {
2698          /* assume full copy single-mip images use shader read access */
2699          t.oldLayout = VK_IMAGE_LAYOUT_GENERAL;
2700          t.newLayout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
2701          VKSCR(TransitionImageLayoutEXT)(screen->dev, 1, &t);
2702          res->layout = VK_IMAGE_LAYOUT_SHADER_READ_ONLY_OPTIMAL;
2703          /* assume multi-mip where further subdata calls may happen */
2704       }
2705       /* make sure image is marked as having data */
2706       res->valid = true;
2707       return;
2708    }
2709    /* fallback case for per-resource unsupported or device-level unsupported */
2710    u_default_texture_subdata(pctx, pres, level, usage, box, data, stride, layer_stride);
2711 }
2712 
2713 static void
zink_transfer_flush_region(struct pipe_context * pctx,struct pipe_transfer * ptrans,const struct pipe_box * box)2714 zink_transfer_flush_region(struct pipe_context *pctx,
2715                            struct pipe_transfer *ptrans,
2716                            const struct pipe_box *box)
2717 {
2718    struct zink_context *ctx = zink_context(pctx);
2719    struct zink_resource *res = zink_resource(ptrans->resource);
2720    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
2721 
2722    if (trans->base.b.usage & PIPE_MAP_WRITE) {
2723       struct zink_screen *screen = zink_screen(pctx->screen);
2724       struct zink_resource *m = trans->staging_res ? zink_resource(trans->staging_res) :
2725                                                      res;
2726       ASSERTED VkDeviceSize size, src_offset, dst_offset = 0;
2727       if (m->obj->is_buffer) {
2728          size = box->width;
2729          src_offset = box->x + (trans->staging_res ? trans->offset : ptrans->box.x);
2730          dst_offset = box->x + ptrans->box.x;
2731       } else {
2732          size = (VkDeviceSize)box->width * box->height * util_format_get_blocksize(m->base.b.format);
2733          src_offset = trans->offset +
2734                   box->z * trans->depthPitch +
2735                   util_format_get_2d_size(m->base.b.format, trans->base.b.stride, box->y) +
2736                   util_format_get_stride(m->base.b.format, box->x);
2737          assert(src_offset + size <= res->obj->size);
2738       }
2739       if (!m->obj->coherent) {
2740          VkMappedMemoryRange range = zink_resource_init_mem_range(screen, m->obj, m->obj->offset, m->obj->size);
2741          if (VKSCR(FlushMappedMemoryRanges)(screen->dev, 1, &range) != VK_SUCCESS) {
2742             mesa_loge("ZINK: vkFlushMappedMemoryRanges failed");
2743          }
2744       }
2745       if (trans->staging_res) {
2746          struct zink_resource *staging_res = zink_resource(trans->staging_res);
2747 
2748          if (ptrans->resource->target == PIPE_BUFFER)
2749             zink_copy_buffer(ctx, res, staging_res, dst_offset, src_offset, size);
2750          else
2751             zink_transfer_copy_bufimage(ctx, res, staging_res, trans);
2752       }
2753    }
2754 }
2755 
2756 /* used to determine whether to emit a TRANSFER_DST barrier on copies */
2757 bool
zink_resource_copy_box_intersects(struct zink_resource * res,unsigned level,const struct pipe_box * box)2758 zink_resource_copy_box_intersects(struct zink_resource *res, unsigned level, const struct pipe_box *box)
2759 {
2760    /* if there are no valid copy rects tracked, this needs a barrier */
2761    if (!res->obj->copies_valid)
2762       return true;
2763    /* untracked huge miplevel */
2764    if (level >= ARRAY_SIZE(res->obj->copies))
2765       return true;
2766    u_rwlock_rdlock(&res->obj->copy_lock);
2767    struct pipe_box *b = res->obj->copies[level].data;
2768    unsigned num_boxes = util_dynarray_num_elements(&res->obj->copies[level], struct pipe_box);
2769    bool (*intersect)(const struct pipe_box *, const struct pipe_box *);
2770    /* determine intersection function based on dimensionality */
2771    switch (res->base.b.target) {
2772    case PIPE_BUFFER:
2773    case PIPE_TEXTURE_1D:
2774       intersect = u_box_test_intersection_1d;
2775       break;
2776 
2777    case PIPE_TEXTURE_1D_ARRAY:
2778    case PIPE_TEXTURE_2D:
2779       intersect = u_box_test_intersection_2d;
2780       break;
2781 
2782    default:
2783       intersect = u_box_test_intersection_3d;
2784       break;
2785    }
2786    /* if any of the tracked boxes intersect with this one, a barrier is needed */
2787    bool ret = false;
2788    for (unsigned i = 0; i < num_boxes; i++) {
2789       if (intersect(box, b + i)) {
2790          ret = true;
2791          break;
2792       }
2793    }
2794    u_rwlock_rdunlock(&res->obj->copy_lock);
2795    /* no intersection = no barrier */
2796    return ret;
2797 }
2798 
2799 /* track a new region for TRANSFER_DST barrier emission */
2800 void
zink_resource_copy_box_add(struct zink_context * ctx,struct zink_resource * res,unsigned level,const struct pipe_box * box)2801 zink_resource_copy_box_add(struct zink_context *ctx, struct zink_resource *res, unsigned level, const struct pipe_box *box)
2802 {
2803    u_rwlock_wrlock(&res->obj->copy_lock);
2804    if (res->obj->copies_valid) {
2805       struct pipe_box *b = res->obj->copies[level].data;
2806       unsigned num_boxes = util_dynarray_num_elements(&res->obj->copies[level], struct pipe_box);
2807       for (unsigned i = 0; i < num_boxes; i++) {
2808          switch (res->base.b.target) {
2809          case PIPE_BUFFER:
2810          case PIPE_TEXTURE_1D:
2811             /* no-op included region */
2812             if (b[i].x <= box->x && b[i].x + b[i].width >= box->x + box->width)
2813                goto out;
2814 
2815             /* try to merge adjacent regions */
2816             if (b[i].x == box->x + box->width) {
2817                b[i].x -= box->width;
2818                b[i].width += box->width;
2819                goto out;
2820             }
2821             if (b[i].x + b[i].width == box->x) {
2822                b[i].width += box->width;
2823                goto out;
2824             }
2825 
2826             /* try to merge into region */
2827             if (box->x <= b[i].x && box->x + box->width >= b[i].x + b[i].width) {
2828                *b = *box;
2829                goto out;
2830             }
2831             break;
2832 
2833          case PIPE_TEXTURE_1D_ARRAY:
2834          case PIPE_TEXTURE_2D:
2835             /* no-op included region */
2836             if (b[i].x <= box->x && b[i].x + b[i].width >= box->x + box->width &&
2837                 b[i].y <= box->y && b[i].y + b[i].height >= box->y + box->height)
2838                goto out;
2839 
2840             /* try to merge adjacent regions */
2841             if (b[i].y == box->y && b[i].height == box->height) {
2842                if (b[i].x == box->x + box->width) {
2843                   b[i].x -= box->width;
2844                   b[i].width += box->width;
2845                   goto out;
2846                }
2847                if (b[i].x + b[i].width == box->x) {
2848                   b[i].width += box->width;
2849                   goto out;
2850                }
2851             } else if (b[i].x == box->x && b[i].width == box->width) {
2852                if (b[i].y == box->y + box->height) {
2853                   b[i].y -= box->height;
2854                   b[i].height += box->height;
2855                   goto out;
2856                }
2857                if (b[i].y + b[i].height == box->y) {
2858                   b[i].height += box->height;
2859                   goto out;
2860                }
2861             }
2862 
2863             /* try to merge into region */
2864             if (box->x <= b[i].x && box->x + box->width >= b[i].x + b[i].width &&
2865                 box->y <= b[i].y && box->y + box->height >= b[i].y + b[i].height) {
2866                *b = *box;
2867                goto out;
2868             }
2869             break;
2870 
2871          default:
2872             /* no-op included region */
2873             if (b[i].x <= box->x && b[i].x + b[i].width >= box->x + box->width &&
2874                 b[i].y <= box->y && b[i].y + b[i].height >= box->y + box->height &&
2875                 b[i].z <= box->z && b[i].z + b[i].depth >= box->z + box->depth)
2876                goto out;
2877 
2878                /* try to merge adjacent regions */
2879             if (b[i].z == box->z && b[i].depth == box->depth) {
2880                if (b[i].y == box->y && b[i].height == box->height) {
2881                   if (b[i].x == box->x + box->width) {
2882                      b[i].x -= box->width;
2883                      b[i].width += box->width;
2884                      goto out;
2885                   }
2886                   if (b[i].x + b[i].width == box->x) {
2887                      b[i].width += box->width;
2888                      goto out;
2889                   }
2890                } else if (b[i].x == box->x && b[i].width == box->width) {
2891                   if (b[i].y == box->y + box->height) {
2892                      b[i].y -= box->height;
2893                      b[i].height += box->height;
2894                      goto out;
2895                   }
2896                   if (b[i].y + b[i].height == box->y) {
2897                      b[i].height += box->height;
2898                      goto out;
2899                   }
2900                }
2901             } else if (b[i].x == box->x && b[i].width == box->width) {
2902                if (b[i].y == box->y && b[i].height == box->height) {
2903                   if (b[i].z == box->z + box->depth) {
2904                      b[i].z -= box->depth;
2905                      b[i].depth += box->depth;
2906                      goto out;
2907                   }
2908                   if (b[i].z + b[i].depth == box->z) {
2909                      b[i].depth += box->depth;
2910                      goto out;
2911                   }
2912                } else if (b[i].z == box->z && b[i].depth == box->depth) {
2913                   if (b[i].y == box->y + box->height) {
2914                      b[i].y -= box->height;
2915                      b[i].height += box->height;
2916                      goto out;
2917                   }
2918                   if (b[i].y + b[i].height == box->y) {
2919                      b[i].height += box->height;
2920                      goto out;
2921                   }
2922                }
2923             } else if (b[i].y == box->y && b[i].height == box->height) {
2924                if (b[i].z == box->z && b[i].depth == box->depth) {
2925                   if (b[i].x == box->x + box->width) {
2926                      b[i].x -= box->width;
2927                      b[i].width += box->width;
2928                      goto out;
2929                   }
2930                   if (b[i].x + b[i].width == box->x) {
2931                      b[i].width += box->width;
2932                      goto out;
2933                   }
2934                } else if (b[i].x == box->x && b[i].width == box->width) {
2935                   if (b[i].z == box->z + box->depth) {
2936                      b[i].z -= box->depth;
2937                      b[i].depth += box->depth;
2938                      goto out;
2939                   }
2940                   if (b[i].z + b[i].depth == box->z) {
2941                      b[i].depth += box->depth;
2942                      goto out;
2943                   }
2944                }
2945             }
2946 
2947             /* try to merge into region */
2948             if (box->x <= b[i].x && box->x + box->width >= b[i].x + b[i].width &&
2949                 box->y <= b[i].y && box->y + box->height >= b[i].y + b[i].height &&
2950                 box->z <= b[i].z && box->z + box->depth >= b[i].z + b[i].depth)
2951                goto out;
2952 
2953             break;
2954          }
2955       }
2956    }
2957    util_dynarray_append(&res->obj->copies[level], struct pipe_box, *box);
2958    if (!res->copies_warned && util_dynarray_num_elements(&res->obj->copies[level], struct pipe_box) > 100) {
2959       perf_debug(ctx, "zink: PERF WARNING! > 100 copy boxes detected for %p\n", res);
2960       mesa_logw("zink: PERF WARNING! > 100 copy boxes detected for %p\n", res);
2961       res->copies_warned = true;
2962    }
2963    res->obj->copies_valid = true;
2964 out:
2965    u_rwlock_wrunlock(&res->obj->copy_lock);
2966 }
2967 
2968 void
zink_resource_copies_reset(struct zink_resource * res)2969 zink_resource_copies_reset(struct zink_resource *res)
2970 {
2971    if (!res->obj->copies_valid)
2972       return;
2973    u_rwlock_wrlock(&res->obj->copy_lock);
2974    unsigned max_level = res->base.b.target == PIPE_BUFFER ? 1 : (res->base.b.last_level + 1);
2975    if (res->base.b.target == PIPE_BUFFER) {
2976       /* flush transfer regions back to valid range on reset */
2977       struct pipe_box *b = res->obj->copies[0].data;
2978       unsigned num_boxes = util_dynarray_num_elements(&res->obj->copies[0], struct pipe_box);
2979       for (unsigned i = 0; i < num_boxes; i++)
2980          util_range_add(&res->base.b, &res->valid_buffer_range, b[i].x, b[i].x + b[i].width);
2981    }
2982    for (unsigned i = 0; i < max_level; i++)
2983       util_dynarray_clear(&res->obj->copies[i]);
2984    res->obj->copies_valid = false;
2985    res->obj->copies_need_reset = false;
2986    u_rwlock_wrunlock(&res->obj->copy_lock);
2987 }
2988 
2989 static void
transfer_unmap(struct pipe_context * pctx,struct pipe_transfer * ptrans)2990 transfer_unmap(struct pipe_context *pctx, struct pipe_transfer *ptrans)
2991 {
2992    struct zink_context *ctx = zink_context(pctx);
2993    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
2994 
2995    if (!(trans->base.b.usage & (PIPE_MAP_FLUSH_EXPLICIT | PIPE_MAP_COHERENT))) {
2996       /* flush_region is relative to the mapped region: use only the extents */
2997       struct pipe_box box = ptrans->box;
2998       box.x = box.y = box.z = 0;
2999       zink_transfer_flush_region(pctx, ptrans, &box);
3000    }
3001 
3002    if (trans->staging_res)
3003       pipe_resource_reference(&trans->staging_res, NULL);
3004    pipe_resource_reference(&trans->base.b.resource, NULL);
3005 
3006    destroy_transfer(ctx, trans);
3007 }
3008 
3009 static void
do_transfer_unmap(struct zink_screen * screen,struct zink_transfer * trans)3010 do_transfer_unmap(struct zink_screen *screen, struct zink_transfer *trans)
3011 {
3012    struct zink_resource *res = zink_resource(trans->staging_res);
3013    if (!res)
3014       res = zink_resource(trans->base.b.resource);
3015    unmap_resource(screen, res);
3016 }
3017 
3018 void
zink_screen_buffer_unmap(struct pipe_screen * pscreen,struct pipe_transfer * ptrans)3019 zink_screen_buffer_unmap(struct pipe_screen *pscreen, struct pipe_transfer *ptrans)
3020 {
3021    struct zink_screen *screen = zink_screen(pscreen);
3022    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
3023    if (trans->base.b.usage & PIPE_MAP_ONCE && !trans->staging_res)
3024       do_transfer_unmap(screen, trans);
3025    transfer_unmap(NULL, ptrans);
3026 }
3027 
3028 static void
zink_buffer_unmap(struct pipe_context * pctx,struct pipe_transfer * ptrans)3029 zink_buffer_unmap(struct pipe_context *pctx, struct pipe_transfer *ptrans)
3030 {
3031    struct zink_screen *screen = zink_screen(pctx->screen);
3032    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
3033    if (trans->base.b.usage & PIPE_MAP_ONCE && !trans->staging_res)
3034       do_transfer_unmap(screen, trans);
3035    transfer_unmap(pctx, ptrans);
3036 }
3037 
3038 static void
zink_image_unmap(struct pipe_context * pctx,struct pipe_transfer * ptrans)3039 zink_image_unmap(struct pipe_context *pctx, struct pipe_transfer *ptrans)
3040 {
3041    struct zink_screen *screen = zink_screen(pctx->screen);
3042    struct zink_transfer *trans = (struct zink_transfer *)ptrans;
3043    if (sizeof(void*) == 4)
3044       do_transfer_unmap(screen, trans);
3045    transfer_unmap(pctx, ptrans);
3046 }
3047 
3048 static void
zink_buffer_subdata(struct pipe_context * ctx,struct pipe_resource * buffer,unsigned usage,unsigned offset,unsigned size,const void * data)3049 zink_buffer_subdata(struct pipe_context *ctx, struct pipe_resource *buffer,
3050                     unsigned usage, unsigned offset, unsigned size, const void *data)
3051 {
3052    struct pipe_transfer *transfer = NULL;
3053    struct pipe_box box;
3054    uint8_t *map = NULL;
3055 
3056    usage |= PIPE_MAP_WRITE;
3057 
3058    if (!(usage & PIPE_MAP_DIRECTLY))
3059       usage |= PIPE_MAP_DISCARD_RANGE;
3060 
3061    u_box_1d(offset, size, &box);
3062    map = zink_buffer_map(ctx, buffer, 0, usage, &box, &transfer);
3063    if (!map)
3064       return;
3065 
3066    memcpy(map, data, size);
3067    zink_buffer_unmap(ctx, transfer);
3068 }
3069 
3070 static struct pipe_resource *
zink_resource_get_separate_stencil(struct pipe_resource * pres)3071 zink_resource_get_separate_stencil(struct pipe_resource *pres)
3072 {
3073    /* For packed depth-stencil, we treat depth as the primary resource
3074     * and store S8 as the "second plane" resource.
3075     */
3076    if (pres->next && pres->next->format == PIPE_FORMAT_S8_UINT)
3077       return pres->next;
3078 
3079    return NULL;
3080 
3081 }
3082 
3083 static bool
resource_object_add_bind(struct zink_context * ctx,struct zink_resource * res,unsigned bind)3084 resource_object_add_bind(struct zink_context *ctx, struct zink_resource *res, unsigned bind)
3085 {
3086    /* base resource already has the cap */
3087    if (res->base.b.bind & bind)
3088       return true;
3089    if (res->obj->is_buffer) {
3090       unreachable("zink: all buffers should have this bit");
3091       return true;
3092    }
3093    assert(!res->obj->dt);
3094    zink_fb_clears_apply_region(ctx, &res->base.b, (struct u_rect){0, res->base.b.width0, 0, res->base.b.height0});
3095    bool ret = add_resource_bind(ctx, res, bind);
3096    if (ret)
3097       zink_resource_rebind(ctx, res);
3098 
3099    return ret;
3100 }
3101 
3102 bool
zink_resource_object_init_storage(struct zink_context * ctx,struct zink_resource * res)3103 zink_resource_object_init_storage(struct zink_context *ctx, struct zink_resource *res)
3104 {
3105    return resource_object_add_bind(ctx, res, PIPE_BIND_SHADER_IMAGE);
3106 }
3107 
3108 bool
zink_resource_object_init_mutable(struct zink_context * ctx,struct zink_resource * res)3109 zink_resource_object_init_mutable(struct zink_context *ctx, struct zink_resource *res)
3110 {
3111    return resource_object_add_bind(ctx, res, ZINK_BIND_MUTABLE);
3112 }
3113 
3114 VkDeviceAddress
zink_resource_get_address(struct zink_screen * screen,struct zink_resource * res)3115 zink_resource_get_address(struct zink_screen *screen, struct zink_resource *res)
3116 {
3117    assert(res->obj->is_buffer);
3118    if (!res->obj->bda) {
3119       VkBufferDeviceAddressInfo info = {
3120          VK_STRUCTURE_TYPE_BUFFER_DEVICE_ADDRESS_INFO,
3121          NULL,
3122          res->obj->buffer
3123       };
3124       res->obj->bda = VKSCR(GetBufferDeviceAddress)(screen->dev, &info);
3125    }
3126    return res->obj->bda;
3127 }
3128 
3129 void
zink_resource_setup_transfer_layouts(struct zink_context * ctx,struct zink_resource * src,struct zink_resource * dst)3130 zink_resource_setup_transfer_layouts(struct zink_context *ctx, struct zink_resource *src, struct zink_resource *dst)
3131 {
3132    if (src == dst) {
3133       /* The Vulkan 1.1 specification says the following about valid usage
3134        * of vkCmdBlitImage:
3135        *
3136        * "srcImageLayout must be VK_IMAGE_LAYOUT_SHARED_PRESENT_KHR,
3137        *  VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL or VK_IMAGE_LAYOUT_GENERAL"
3138        *
3139        * and:
3140        *
3141        * "dstImageLayout must be VK_IMAGE_LAYOUT_SHARED_PRESENT_KHR,
3142        *  VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL or VK_IMAGE_LAYOUT_GENERAL"
3143        *
3144        * Since we cant have the same image in two states at the same time,
3145        * we're effectively left with VK_IMAGE_LAYOUT_SHARED_PRESENT_KHR or
3146        * VK_IMAGE_LAYOUT_GENERAL. And since this isn't a present-related
3147        * operation, VK_IMAGE_LAYOUT_GENERAL seems most appropriate.
3148        */
3149       zink_screen(ctx->base.screen)->image_barrier(ctx, src,
3150                                   VK_IMAGE_LAYOUT_GENERAL,
3151                                   VK_ACCESS_TRANSFER_READ_BIT | VK_ACCESS_TRANSFER_WRITE_BIT,
3152                                   VK_PIPELINE_STAGE_TRANSFER_BIT);
3153    } else {
3154       zink_screen(ctx->base.screen)->image_barrier(ctx, src,
3155                                   VK_IMAGE_LAYOUT_TRANSFER_SRC_OPTIMAL,
3156                                   VK_ACCESS_TRANSFER_READ_BIT,
3157                                   VK_PIPELINE_STAGE_TRANSFER_BIT);
3158 
3159       zink_screen(ctx->base.screen)->image_barrier(ctx, dst,
3160                                   VK_IMAGE_LAYOUT_TRANSFER_DST_OPTIMAL,
3161                                   VK_ACCESS_TRANSFER_WRITE_BIT,
3162                                   VK_PIPELINE_STAGE_TRANSFER_BIT);
3163    }
3164 }
3165 
3166 void
zink_get_depth_stencil_resources(struct pipe_resource * res,struct zink_resource ** out_z,struct zink_resource ** out_s)3167 zink_get_depth_stencil_resources(struct pipe_resource *res,
3168                                  struct zink_resource **out_z,
3169                                  struct zink_resource **out_s)
3170 {
3171    if (!res) {
3172       if (out_z) *out_z = NULL;
3173       if (out_s) *out_s = NULL;
3174       return;
3175    }
3176 
3177    if (res->format != PIPE_FORMAT_S8_UINT) {
3178       if (out_z) *out_z = zink_resource(res);
3179       if (out_s) *out_s = zink_resource(zink_resource_get_separate_stencil(res));
3180    } else {
3181       if (out_z) *out_z = NULL;
3182       if (out_s) *out_s = zink_resource(res);
3183    }
3184 }
3185 
3186 static void
zink_resource_set_separate_stencil(struct pipe_resource * pres,struct pipe_resource * stencil)3187 zink_resource_set_separate_stencil(struct pipe_resource *pres,
3188                                    struct pipe_resource *stencil)
3189 {
3190    assert(util_format_has_depth(util_format_description(pres->format)));
3191    pipe_resource_reference(&pres->next, stencil);
3192 }
3193 
3194 static enum pipe_format
zink_resource_get_internal_format(struct pipe_resource * pres)3195 zink_resource_get_internal_format(struct pipe_resource *pres)
3196 {
3197    struct zink_resource *res = zink_resource(pres);
3198    return res->internal_format;
3199 }
3200 
3201 static const struct u_transfer_vtbl transfer_vtbl = {
3202    .resource_create       = zink_resource_create,
3203    .resource_destroy      = zink_resource_destroy,
3204    .transfer_map          = zink_image_map,
3205    .transfer_unmap        = zink_image_unmap,
3206    .transfer_flush_region = zink_transfer_flush_region,
3207    .get_internal_format   = zink_resource_get_internal_format,
3208    .set_stencil           = zink_resource_set_separate_stencil,
3209    .get_stencil           = zink_resource_get_separate_stencil,
3210 };
3211 
3212 bool
zink_screen_resource_init(struct pipe_screen * pscreen)3213 zink_screen_resource_init(struct pipe_screen *pscreen)
3214 {
3215    struct zink_screen *screen = zink_screen(pscreen);
3216    pscreen->resource_create = u_transfer_helper_resource_create;
3217    pscreen->resource_create_with_modifiers = zink_resource_create_with_modifiers;
3218    pscreen->resource_create_drawable = zink_resource_create_drawable;
3219    pscreen->resource_destroy = u_transfer_helper_resource_destroy;
3220    pscreen->transfer_helper = u_transfer_helper_create(&transfer_vtbl,
3221       U_TRANSFER_HELPER_SEPARATE_Z32S8 | U_TRANSFER_HELPER_SEPARATE_STENCIL |
3222       U_TRANSFER_HELPER_INTERLEAVE_IN_PLACE |
3223       U_TRANSFER_HELPER_MSAA_MAP |
3224       (!screen->have_D24_UNORM_S8_UINT ? U_TRANSFER_HELPER_Z24_IN_Z32F : 0));
3225 
3226    if (screen->info.have_KHR_external_memory_fd || screen->info.have_KHR_external_memory_win32) {
3227       pscreen->resource_get_handle = zink_resource_get_handle;
3228       pscreen->resource_from_handle = zink_resource_from_handle;
3229    }
3230    if (screen->info.have_EXT_external_memory_host) {
3231       pscreen->resource_from_user_memory = zink_resource_from_user_memory;
3232    }
3233    if (screen->instance_info.have_KHR_external_memory_capabilities) {
3234       pscreen->memobj_create_from_handle = zink_memobj_create_from_handle;
3235       pscreen->memobj_destroy = zink_memobj_destroy;
3236       pscreen->resource_from_memobj = zink_resource_from_memobj;
3237    }
3238    pscreen->resource_get_param = zink_resource_get_param;
3239    return true;
3240 }
3241 
3242 void
zink_context_resource_init(struct pipe_context * pctx)3243 zink_context_resource_init(struct pipe_context *pctx)
3244 {
3245    pctx->buffer_map = zink_buffer_map;
3246    pctx->buffer_unmap = zink_buffer_unmap;
3247    pctx->texture_map = u_transfer_helper_transfer_map;
3248    pctx->texture_unmap = u_transfer_helper_transfer_unmap;
3249 
3250    pctx->transfer_flush_region = u_transfer_helper_transfer_flush_region;
3251    pctx->buffer_subdata = zink_buffer_subdata;
3252    pctx->texture_subdata = zink_image_subdata;
3253    pctx->invalidate_resource = zink_resource_invalidate;
3254 }
3255