1 /*
2 * Copyright © 2016 Red Hat
3 * based on intel anv code:
4 * Copyright © 2015 Intel Corporation
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a
7 * copy of this software and associated documentation files (the "Software"),
8 * to deal in the Software without restriction, including without limitation
9 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
10 * and/or sell copies of the Software, and to permit persons to whom the
11 * Software is furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice (including the next
14 * paragraph) shall be included in all copies or substantial portions of the
15 * Software.
16 *
17 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
18 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
19 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
20 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
21 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
22 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
23 * IN THE SOFTWARE.
24 */
25
26 #include "radv_meta.h"
27
28 #include <fcntl.h>
29 #include <limits.h>
30 #include <pwd.h>
31 #include <sys/stat.h>
32
33 void
radv_meta_save(struct radv_meta_saved_state * state,const struct radv_cmd_buffer * cmd_buffer,uint32_t dynamic_mask)34 radv_meta_save(struct radv_meta_saved_state *state,
35 const struct radv_cmd_buffer *cmd_buffer,
36 uint32_t dynamic_mask)
37 {
38 state->old_pipeline = cmd_buffer->state.pipeline;
39 state->old_descriptor_set0 = cmd_buffer->state.descriptors[0];
40 memcpy(state->old_vertex_bindings, cmd_buffer->state.vertex_bindings,
41 sizeof(state->old_vertex_bindings));
42
43 state->dynamic_mask = dynamic_mask;
44 radv_dynamic_state_copy(&state->dynamic, &cmd_buffer->state.dynamic,
45 dynamic_mask);
46
47 memcpy(state->push_constants, cmd_buffer->push_constants, MAX_PUSH_CONSTANTS_SIZE);
48 }
49
50 void
radv_meta_restore(const struct radv_meta_saved_state * state,struct radv_cmd_buffer * cmd_buffer)51 radv_meta_restore(const struct radv_meta_saved_state *state,
52 struct radv_cmd_buffer *cmd_buffer)
53 {
54 cmd_buffer->state.pipeline = state->old_pipeline;
55 radv_bind_descriptor_set(cmd_buffer, state->old_descriptor_set0, 0);
56 memcpy(cmd_buffer->state.vertex_bindings, state->old_vertex_bindings,
57 sizeof(state->old_vertex_bindings));
58
59 cmd_buffer->state.vb_dirty |= (1 << RADV_META_VERTEX_BINDING_COUNT) - 1;
60 cmd_buffer->state.dirty |= RADV_CMD_DIRTY_PIPELINE;
61
62 radv_dynamic_state_copy(&cmd_buffer->state.dynamic, &state->dynamic,
63 state->dynamic_mask);
64 cmd_buffer->state.dirty |= state->dynamic_mask;
65
66 memcpy(cmd_buffer->push_constants, state->push_constants, MAX_PUSH_CONSTANTS_SIZE);
67 cmd_buffer->push_constant_stages |= VK_SHADER_STAGE_ALL_GRAPHICS | VK_SHADER_STAGE_COMPUTE_BIT;
68 }
69
70 void
radv_meta_save_pass(struct radv_meta_saved_pass_state * state,const struct radv_cmd_buffer * cmd_buffer)71 radv_meta_save_pass(struct radv_meta_saved_pass_state *state,
72 const struct radv_cmd_buffer *cmd_buffer)
73 {
74 state->pass = cmd_buffer->state.pass;
75 state->subpass = cmd_buffer->state.subpass;
76 state->framebuffer = cmd_buffer->state.framebuffer;
77 state->attachments = cmd_buffer->state.attachments;
78 state->render_area = cmd_buffer->state.render_area;
79 }
80
81 void
radv_meta_restore_pass(const struct radv_meta_saved_pass_state * state,struct radv_cmd_buffer * cmd_buffer)82 radv_meta_restore_pass(const struct radv_meta_saved_pass_state *state,
83 struct radv_cmd_buffer *cmd_buffer)
84 {
85 cmd_buffer->state.pass = state->pass;
86 cmd_buffer->state.subpass = state->subpass;
87 cmd_buffer->state.framebuffer = state->framebuffer;
88 cmd_buffer->state.attachments = state->attachments;
89 cmd_buffer->state.render_area = state->render_area;
90 if (state->subpass)
91 radv_emit_framebuffer_state(cmd_buffer);
92 }
93
94 void
radv_meta_save_compute(struct radv_meta_saved_compute_state * state,const struct radv_cmd_buffer * cmd_buffer,unsigned push_constant_size)95 radv_meta_save_compute(struct radv_meta_saved_compute_state *state,
96 const struct radv_cmd_buffer *cmd_buffer,
97 unsigned push_constant_size)
98 {
99 state->old_pipeline = cmd_buffer->state.compute_pipeline;
100 state->old_descriptor_set0 = cmd_buffer->state.descriptors[0];
101
102 if (push_constant_size)
103 memcpy(state->push_constants, cmd_buffer->push_constants, push_constant_size);
104 }
105
106 void
radv_meta_restore_compute(const struct radv_meta_saved_compute_state * state,struct radv_cmd_buffer * cmd_buffer,unsigned push_constant_size)107 radv_meta_restore_compute(const struct radv_meta_saved_compute_state *state,
108 struct radv_cmd_buffer *cmd_buffer,
109 unsigned push_constant_size)
110 {
111 radv_CmdBindPipeline(radv_cmd_buffer_to_handle(cmd_buffer), VK_PIPELINE_BIND_POINT_COMPUTE,
112 radv_pipeline_to_handle(state->old_pipeline));
113 radv_bind_descriptor_set(cmd_buffer, state->old_descriptor_set0, 0);
114
115 if (push_constant_size) {
116 memcpy(cmd_buffer->push_constants, state->push_constants, push_constant_size);
117 cmd_buffer->push_constant_stages |= VK_SHADER_STAGE_COMPUTE_BIT;
118 }
119 }
120
121 VkImageViewType
radv_meta_get_view_type(const struct radv_image * image)122 radv_meta_get_view_type(const struct radv_image *image)
123 {
124 switch (image->type) {
125 case VK_IMAGE_TYPE_1D: return VK_IMAGE_VIEW_TYPE_1D;
126 case VK_IMAGE_TYPE_2D: return VK_IMAGE_VIEW_TYPE_2D;
127 case VK_IMAGE_TYPE_3D: return VK_IMAGE_VIEW_TYPE_3D;
128 default:
129 unreachable("bad VkImageViewType");
130 }
131 }
132
133 /**
134 * When creating a destination VkImageView, this function provides the needed
135 * VkImageViewCreateInfo::subresourceRange::baseArrayLayer.
136 */
137 uint32_t
radv_meta_get_iview_layer(const struct radv_image * dest_image,const VkImageSubresourceLayers * dest_subresource,const VkOffset3D * dest_offset)138 radv_meta_get_iview_layer(const struct radv_image *dest_image,
139 const VkImageSubresourceLayers *dest_subresource,
140 const VkOffset3D *dest_offset)
141 {
142 switch (dest_image->type) {
143 case VK_IMAGE_TYPE_1D:
144 case VK_IMAGE_TYPE_2D:
145 return dest_subresource->baseArrayLayer;
146 case VK_IMAGE_TYPE_3D:
147 /* HACK: Vulkan does not allow attaching a 3D image to a framebuffer,
148 * but meta does it anyway. When doing so, we translate the
149 * destination's z offset into an array offset.
150 */
151 return dest_offset->z;
152 default:
153 assert(!"bad VkImageType");
154 return 0;
155 }
156 }
157
158 static void *
meta_alloc(void * _device,size_t size,size_t alignment,VkSystemAllocationScope allocationScope)159 meta_alloc(void* _device, size_t size, size_t alignment,
160 VkSystemAllocationScope allocationScope)
161 {
162 struct radv_device *device = _device;
163 return device->alloc.pfnAllocation(device->alloc.pUserData, size, alignment,
164 VK_SYSTEM_ALLOCATION_SCOPE_DEVICE);
165 }
166
167 static void *
meta_realloc(void * _device,void * original,size_t size,size_t alignment,VkSystemAllocationScope allocationScope)168 meta_realloc(void* _device, void *original, size_t size, size_t alignment,
169 VkSystemAllocationScope allocationScope)
170 {
171 struct radv_device *device = _device;
172 return device->alloc.pfnReallocation(device->alloc.pUserData, original,
173 size, alignment,
174 VK_SYSTEM_ALLOCATION_SCOPE_DEVICE);
175 }
176
177 static void
meta_free(void * _device,void * data)178 meta_free(void* _device, void *data)
179 {
180 struct radv_device *device = _device;
181 return device->alloc.pfnFree(device->alloc.pUserData, data);
182 }
183
184 static bool
radv_builtin_cache_path(char * path)185 radv_builtin_cache_path(char *path)
186 {
187 char *xdg_cache_home = getenv("XDG_CACHE_HOME");
188 const char *suffix = "/radv_builtin_shaders";
189 const char *suffix2 = "/.cache/radv_builtin_shaders";
190 struct passwd pwd, *result;
191 char path2[PATH_MAX + 1]; /* PATH_MAX is not a real max,but suffices here. */
192
193 if (xdg_cache_home) {
194
195 if (strlen(xdg_cache_home) + strlen(suffix) > PATH_MAX)
196 return false;
197
198 strcpy(path, xdg_cache_home);
199 strcat(path, suffix);
200 return true;
201 }
202
203 getpwuid_r(getuid(), &pwd, path2, PATH_MAX - strlen(suffix2), &result);
204 if (!result)
205 return false;
206
207 strcpy(path, pwd.pw_dir);
208 strcat(path, "/.cache");
209 mkdir(path, 0755);
210
211 strcat(path, suffix);
212 return true;
213 }
214
215 static void
radv_load_meta_pipeline(struct radv_device * device)216 radv_load_meta_pipeline(struct radv_device *device)
217 {
218 char path[PATH_MAX + 1];
219 struct stat st;
220 void *data = NULL;
221
222 if (!radv_builtin_cache_path(path))
223 return;
224
225 int fd = open(path, O_RDONLY);
226 if (fd < 0)
227 return;
228 if (fstat(fd, &st))
229 goto fail;
230 data = malloc(st.st_size);
231 if (!data)
232 goto fail;
233 if(read(fd, data, st.st_size) == -1)
234 goto fail;
235
236 radv_pipeline_cache_load(&device->meta_state.cache, data, st.st_size);
237 fail:
238 free(data);
239 close(fd);
240 }
241
242 static void
radv_store_meta_pipeline(struct radv_device * device)243 radv_store_meta_pipeline(struct radv_device *device)
244 {
245 char path[PATH_MAX + 1], path2[PATH_MAX + 7];
246 size_t size;
247 void *data = NULL;
248
249 if (!device->meta_state.cache.modified)
250 return;
251
252 if (radv_GetPipelineCacheData(radv_device_to_handle(device),
253 radv_pipeline_cache_to_handle(&device->meta_state.cache),
254 &size, NULL))
255 return;
256
257 if (!radv_builtin_cache_path(path))
258 return;
259
260 strcpy(path2, path);
261 strcat(path2, "XXXXXX");
262 int fd = mkstemp(path2);//open(path, O_WRONLY | O_CREAT, 0600);
263 if (fd < 0)
264 return;
265 data = malloc(size);
266 if (!data)
267 goto fail;
268
269 if (radv_GetPipelineCacheData(radv_device_to_handle(device),
270 radv_pipeline_cache_to_handle(&device->meta_state.cache),
271 &size, data))
272 goto fail;
273 if(write(fd, data, size) == -1)
274 goto fail;
275
276 rename(path2, path);
277 fail:
278 free(data);
279 close(fd);
280 unlink(path2);
281 }
282
283 VkResult
radv_device_init_meta(struct radv_device * device)284 radv_device_init_meta(struct radv_device *device)
285 {
286 VkResult result;
287
288 device->meta_state.alloc = (VkAllocationCallbacks) {
289 .pUserData = device,
290 .pfnAllocation = meta_alloc,
291 .pfnReallocation = meta_realloc,
292 .pfnFree = meta_free,
293 };
294
295 device->meta_state.cache.alloc = device->meta_state.alloc;
296 radv_pipeline_cache_init(&device->meta_state.cache, device);
297 radv_load_meta_pipeline(device);
298
299 result = radv_device_init_meta_clear_state(device);
300 if (result != VK_SUCCESS)
301 goto fail_clear;
302
303 result = radv_device_init_meta_resolve_state(device);
304 if (result != VK_SUCCESS)
305 goto fail_resolve;
306
307 result = radv_device_init_meta_blit_state(device);
308 if (result != VK_SUCCESS)
309 goto fail_blit;
310
311 result = radv_device_init_meta_blit2d_state(device);
312 if (result != VK_SUCCESS)
313 goto fail_blit2d;
314
315 result = radv_device_init_meta_bufimage_state(device);
316 if (result != VK_SUCCESS)
317 goto fail_bufimage;
318
319 result = radv_device_init_meta_depth_decomp_state(device);
320 if (result != VK_SUCCESS)
321 goto fail_depth_decomp;
322
323 result = radv_device_init_meta_buffer_state(device);
324 if (result != VK_SUCCESS)
325 goto fail_buffer;
326
327 result = radv_device_init_meta_fast_clear_flush_state(device);
328 if (result != VK_SUCCESS)
329 goto fail_fast_clear;
330
331 result = radv_device_init_meta_resolve_compute_state(device);
332 if (result != VK_SUCCESS)
333 goto fail_resolve_compute;
334 return VK_SUCCESS;
335
336 fail_resolve_compute:
337 radv_device_finish_meta_fast_clear_flush_state(device);
338 fail_fast_clear:
339 radv_device_finish_meta_buffer_state(device);
340 fail_buffer:
341 radv_device_finish_meta_depth_decomp_state(device);
342 fail_depth_decomp:
343 radv_device_finish_meta_bufimage_state(device);
344 fail_bufimage:
345 radv_device_finish_meta_blit2d_state(device);
346 fail_blit2d:
347 radv_device_finish_meta_blit_state(device);
348 fail_blit:
349 radv_device_finish_meta_resolve_state(device);
350 fail_resolve:
351 radv_device_finish_meta_clear_state(device);
352 fail_clear:
353 radv_pipeline_cache_finish(&device->meta_state.cache);
354 return result;
355 }
356
357 void
radv_device_finish_meta(struct radv_device * device)358 radv_device_finish_meta(struct radv_device *device)
359 {
360 radv_device_finish_meta_clear_state(device);
361 radv_device_finish_meta_resolve_state(device);
362 radv_device_finish_meta_blit_state(device);
363 radv_device_finish_meta_blit2d_state(device);
364 radv_device_finish_meta_bufimage_state(device);
365 radv_device_finish_meta_depth_decomp_state(device);
366 radv_device_finish_meta_buffer_state(device);
367 radv_device_finish_meta_fast_clear_flush_state(device);
368 radv_device_finish_meta_resolve_compute_state(device);
369
370 radv_store_meta_pipeline(device);
371 radv_pipeline_cache_finish(&device->meta_state.cache);
372 }
373
374 /*
375 * The most common meta operations all want to have the viewport
376 * reset and any scissors disabled. The rest of the dynamic state
377 * should have no effect.
378 */
379 void
radv_meta_save_graphics_reset_vport_scissor(struct radv_meta_saved_state * saved_state,struct radv_cmd_buffer * cmd_buffer)380 radv_meta_save_graphics_reset_vport_scissor(struct radv_meta_saved_state *saved_state,
381 struct radv_cmd_buffer *cmd_buffer)
382 {
383 uint32_t dirty_state = (1 << VK_DYNAMIC_STATE_VIEWPORT) | (1 << VK_DYNAMIC_STATE_SCISSOR);
384 radv_meta_save(saved_state, cmd_buffer, dirty_state);
385 cmd_buffer->state.dynamic.viewport.count = 0;
386 cmd_buffer->state.dynamic.scissor.count = 0;
387 cmd_buffer->state.dirty |= dirty_state;
388 }
389