1 /*
2 * Copyright © 2016 Red Hat.
3 * Copyright © 2016 Bas Nieuwenhuizen
4 *
5 * based in part on anv driver which is:
6 * Copyright © 2015 Intel Corporation
7 *
8 * Permission is hereby granted, free of charge, to any person obtaining a
9 * copy of this software and associated documentation files (the "Software"),
10 * to deal in the Software without restriction, including without limitation
11 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
12 * and/or sell copies of the Software, and to permit persons to whom the
13 * Software is furnished to do so, subject to the following conditions:
14 *
15 * The above copyright notice and this permission notice (including the next
16 * paragraph) shall be included in all copies or substantial portions of the
17 * Software.
18 *
19 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
20 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
21 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
22 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
23 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
24 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
25 * IN THE SOFTWARE.
26 */
27
28 #ifndef RADV_SHADER_H
29 #define RADV_SHADER_H
30
31 #include "util/mesa-blake3.h"
32 #include "util/u_math.h"
33 #include "vulkan/runtime/vk_pipeline_cache.h"
34 #include "vulkan/vulkan.h"
35 #include "ac_binary.h"
36 #include "ac_shader_util.h"
37 #include "amd_family.h"
38 #include "radv_constants.h"
39
40 #include "aco_shader_info.h"
41
42 #define RADV_VERT_ATTRIB_MAX MAX2(VERT_ATTRIB_MAX, VERT_ATTRIB_GENERIC0 + MAX_VERTEX_ATTRIBS)
43
44 struct radv_physical_device;
45 struct radv_device;
46 struct radv_pipeline;
47 struct radv_ray_tracing_pipeline;
48 struct radv_shader_args;
49 struct radv_vs_input_state;
50 struct radv_shader_args;
51 struct radv_serialized_shader_arena_block;
52
53 enum {
54 RADV_GRAPHICS_STAGE_BITS =
55 (VK_SHADER_STAGE_ALL_GRAPHICS | VK_SHADER_STAGE_MESH_BIT_EXT | VK_SHADER_STAGE_TASK_BIT_EXT),
56 RADV_RT_STAGE_BITS =
57 (VK_SHADER_STAGE_RAYGEN_BIT_KHR | VK_SHADER_STAGE_ANY_HIT_BIT_KHR | VK_SHADER_STAGE_CLOSEST_HIT_BIT_KHR |
58 VK_SHADER_STAGE_MISS_BIT_KHR | VK_SHADER_STAGE_INTERSECTION_BIT_KHR | VK_SHADER_STAGE_CALLABLE_BIT_KHR)
59 };
60
61 #define RADV_STAGE_MASK ((1 << MESA_VULKAN_SHADER_STAGES) - 1)
62
63 #define radv_foreach_stage(stage, stage_bits) \
64 for (gl_shader_stage stage, __tmp = (gl_shader_stage)((stage_bits)&RADV_STAGE_MASK); stage = ffs(__tmp) - 1, __tmp; \
65 __tmp &= ~(1 << (stage)))
66
67 enum radv_nggc_settings {
68 radv_nggc_none = 0,
69 radv_nggc_front_face = 1 << 0,
70 radv_nggc_back_face = 1 << 1,
71 radv_nggc_face_is_ccw = 1 << 2,
72 radv_nggc_small_primitives = 1 << 3,
73 };
74
75 enum radv_shader_query_state {
76 radv_shader_query_none = 0,
77 radv_shader_query_pipeline_stat = 1 << 0,
78 radv_shader_query_prim_gen = 1 << 1,
79 radv_shader_query_prim_xfb = 1 << 2,
80 };
81
82 enum radv_required_subgroup_size {
83 RADV_REQUIRED_NONE = 0,
84 RADV_REQUIRED_WAVE32 = 1,
85 RADV_REQUIRED_WAVE64 = 2,
86 };
87
88 struct radv_shader_stage_key {
89 uint8_t subgroup_required_size : 2; /* radv_required_subgroup_size */
90 uint8_t subgroup_require_full : 1; /* whether full subgroups are required */
91
92 uint8_t storage_robustness2 : 1;
93 uint8_t uniform_robustness2 : 1;
94 uint8_t vertex_robustness1 : 1;
95
96 uint8_t optimisations_disabled : 1;
97 uint8_t keep_statistic_info : 1;
98
99 /* Shader version (up to 8) to force re-compilation when RADV_BUILD_ID_OVERRIDE is enabled. */
100 uint8_t version : 3;
101
102 /* Whether the mesh shader is used with a task shader. */
103 uint8_t has_task_shader : 1;
104 };
105
106 struct radv_ps_epilog_key {
107 uint32_t spi_shader_col_format;
108 uint32_t spi_shader_z_format;
109
110 /* Bitmasks, each bit represents one of the 8 MRTs. */
111 uint8_t color_is_int8;
112 uint8_t color_is_int10;
113 uint8_t enable_mrt_output_nan_fixup;
114
115 uint32_t colors_written;
116 bool mrt0_is_dual_src;
117 bool export_depth;
118 bool export_stencil;
119 bool export_sample_mask;
120 bool alpha_to_coverage_via_mrtz;
121 };
122
123 struct radv_spirv_to_nir_options {
124 uint32_t lower_view_index_to_zero : 1;
125 uint32_t fix_dual_src_mrt1_export : 1;
126 };
127
128 struct radv_graphics_state_key {
129 uint32_t lib_flags : 4; /* VkGraphicsPipelineLibraryFlagBitsEXT */
130
131 uint32_t has_multiview_view_index : 1;
132 uint32_t adjust_frag_coord_z : 1;
133 uint32_t dynamic_rasterization_samples : 1;
134 uint32_t dynamic_provoking_vtx_mode : 1;
135 uint32_t dynamic_line_rast_mode : 1;
136 uint32_t enable_remove_point_size : 1;
137 uint32_t unknown_rast_prim : 1;
138
139 struct {
140 uint8_t topology;
141 } ia;
142
143 struct {
144 uint32_t instance_rate_inputs;
145 uint32_t instance_rate_divisors[MAX_VERTEX_ATTRIBS];
146 uint8_t vertex_attribute_formats[MAX_VERTEX_ATTRIBS];
147 uint32_t vertex_attribute_bindings[MAX_VERTEX_ATTRIBS];
148 uint32_t vertex_attribute_offsets[MAX_VERTEX_ATTRIBS];
149 uint32_t vertex_attribute_strides[MAX_VERTEX_ATTRIBS];
150 uint8_t vertex_binding_align[MAX_VBS];
151 } vi;
152
153 struct {
154 unsigned patch_control_points;
155 } ts;
156
157 struct {
158 uint32_t provoking_vtx_last : 1;
159 uint32_t line_smooth_enabled : 1;
160 } rs;
161
162 struct {
163 bool sample_shading_enable;
164 bool alpha_to_coverage_via_mrtz; /* GFX11+ */
165 uint8_t rasterization_samples;
166 } ms;
167
168 struct vs {
169 bool has_prolog;
170 } vs;
171
172 struct {
173 struct radv_ps_epilog_key epilog;
174 bool force_vrs_enabled;
175 bool exports_mrtz_via_epilog;
176 bool has_epilog;
177 } ps;
178 };
179
180 struct radv_graphics_pipeline_key {
181 struct radv_graphics_state_key gfx_state;
182
183 struct radv_shader_stage_key stage_info[MESA_VULKAN_SHADER_STAGES];
184 };
185
186 struct radv_nir_compiler_options {
187 bool robust_buffer_access_llvm;
188 bool dump_shader;
189 bool dump_preoptir;
190 bool record_ir;
191 bool record_stats;
192 bool check_ir;
193 uint8_t enable_mrt_output_nan_fixup;
194 bool wgp_mode;
195 const struct radeon_info *info;
196
197 struct {
198 void (*func)(void *private_data, enum aco_compiler_debug_level level, const char *message);
199 void *private_data;
200 } debug;
201 };
202
203 enum radv_ud_index {
204 AC_UD_SCRATCH_RING_OFFSETS = 0,
205 AC_UD_PUSH_CONSTANTS = 1,
206 AC_UD_INLINE_PUSH_CONSTANTS = 2,
207 AC_UD_INDIRECT_DESCRIPTOR_SETS = 3,
208 AC_UD_VIEW_INDEX = 4,
209 AC_UD_STREAMOUT_BUFFERS = 5,
210 AC_UD_SHADER_QUERY_STATE = 6,
211 AC_UD_NGG_PROVOKING_VTX = 7,
212 AC_UD_NGG_CULLING_SETTINGS = 8,
213 AC_UD_NGG_VIEWPORT = 9,
214 AC_UD_VGT_ESGS_RING_ITEMSIZE = 10,
215 AC_UD_FORCE_VRS_RATES = 11,
216 AC_UD_TASK_RING_ENTRY = 12,
217 AC_UD_NUM_VERTS_PER_PRIM = 13,
218 AC_UD_NEXT_STAGE_PC = 14,
219 AC_UD_SHADER_START = 15,
220 AC_UD_VS_VERTEX_BUFFERS = AC_UD_SHADER_START,
221 AC_UD_VS_BASE_VERTEX_START_INSTANCE,
222 AC_UD_VS_PROLOG_INPUTS,
223 AC_UD_VS_MAX_UD,
224 AC_UD_PS_EPILOG_PC,
225 AC_UD_PS_STATE,
226 AC_UD_PS_MAX_UD,
227 AC_UD_CS_GRID_SIZE = AC_UD_SHADER_START,
228 AC_UD_CS_SBT_DESCRIPTORS,
229 AC_UD_CS_RAY_LAUNCH_SIZE_ADDR,
230 AC_UD_CS_RAY_DYNAMIC_CALLABLE_STACK_BASE,
231 AC_UD_CS_TRAVERSAL_SHADER_ADDR,
232 AC_UD_CS_TASK_RING_OFFSETS,
233 AC_UD_CS_TASK_DRAW_ID,
234 AC_UD_CS_TASK_IB,
235 AC_UD_CS_MAX_UD,
236 AC_UD_GS_MAX_UD,
237 AC_UD_TCS_OFFCHIP_LAYOUT = AC_UD_VS_MAX_UD,
238 AC_UD_TCS_EPILOG_PC,
239 AC_UD_TCS_MAX_UD,
240 /* We might not know the previous stage when compiling a geometry shader, so we just
241 * declare both TES and VS user SGPRs.
242 */
243 AC_UD_TES_STATE = AC_UD_VS_MAX_UD,
244 AC_UD_TES_MAX_UD,
245 AC_UD_MAX_UD = AC_UD_CS_MAX_UD,
246 };
247
248 #define SET_SGPR_FIELD(field, value) (((unsigned)(value)&field##__MASK) << field##__SHIFT)
249
250 #define TCS_OFFCHIP_LAYOUT_PATCH_CONTROL_POINTS__SHIFT 0
251 #define TCS_OFFCHIP_LAYOUT_PATCH_CONTROL_POINTS__MASK 0x3f
252 #define TCS_OFFCHIP_LAYOUT_NUM_PATCHES__SHIFT 6
253 #define TCS_OFFCHIP_LAYOUT_NUM_PATCHES__MASK 0x3f
254 #define TCS_OFFCHIP_LAYOUT_LSHS_VERTEX_STRIDE__SHIFT 12
255 #define TCS_OFFCHIP_LAYOUT_LSHS_VERTEX_STRIDE__MASK 0xff /* max 32 * 4 + 1 (to reduce LDS bank conflicts) */
256
257 #define TES_STATE_NUM_PATCHES__SHIFT 0
258 #define TES_STATE_NUM_PATCHES__MASK 0xff
259 #define TES_STATE_TCS_VERTICES_OUT__SHIFT 8
260 #define TES_STATE_TCS_VERTICES_OUT__MASK 0xff
261 #define TES_STATE_NUM_TCS_OUTPUTS__SHIFT 16
262 #define TES_STATE_NUM_TCS_OUTPUTS__MASK 0xff
263
264 #define PS_STATE_NUM_SAMPLES__SHIFT 0
265 #define PS_STATE_NUM_SAMPLES__MASK 0xf
266 #define PS_STATE_LINE_RAST_MODE__SHIFT 4
267 #define PS_STATE_LINE_RAST_MODE__MASK 0x3
268 #define PS_STATE_PS_ITER_MASK__SHIFT 6
269 #define PS_STATE_PS_ITER_MASK__MASK 0xffff
270 #define PS_STATE_RAST_PRIM__SHIFT 22
271 #define PS_STATE_RAST_PRIM__MASK 0x3
272
273 struct radv_streamout_info {
274 uint16_t num_outputs;
275 uint16_t strides[MAX_SO_BUFFERS];
276 uint32_t enabled_stream_buffers_mask;
277 };
278
279 struct radv_userdata_info {
280 int8_t sgpr_idx;
281 uint8_t num_sgprs;
282 };
283
284 struct radv_userdata_locations {
285 struct radv_userdata_info descriptor_sets[MAX_SETS];
286 struct radv_userdata_info shader_data[AC_UD_MAX_UD];
287 uint32_t descriptor_sets_enabled;
288 };
289
290 struct radv_vs_output_info {
291 uint8_t vs_output_param_offset[VARYING_SLOT_MAX];
292 uint8_t clip_dist_mask;
293 uint8_t cull_dist_mask;
294 uint8_t param_exports;
295 uint8_t prim_param_exports;
296 bool writes_pointsize;
297 bool writes_layer;
298 bool writes_layer_per_primitive;
299 bool writes_viewport_index;
300 bool writes_viewport_index_per_primitive;
301 bool writes_primitive_shading_rate;
302 bool writes_primitive_shading_rate_per_primitive;
303 bool export_prim_id;
304 unsigned pos_exports;
305 };
306
307 struct radv_legacy_gs_info {
308 uint32_t vgt_gs_onchip_cntl;
309 uint32_t vgt_gs_max_prims_per_subgroup;
310 uint32_t vgt_esgs_ring_itemsize;
311 uint32_t lds_size;
312 uint32_t esgs_ring_size;
313 uint32_t gsvs_ring_size;
314 };
315
316 struct gfx10_ngg_info {
317 uint16_t ngg_emit_size; /* in dwords */
318 uint32_t hw_max_esverts;
319 uint32_t max_gsprims;
320 uint32_t max_out_verts;
321 uint32_t prim_amp_factor;
322 uint32_t vgt_esgs_ring_itemsize;
323 uint32_t esgs_ring_size;
324 uint32_t scratch_lds_base;
325 uint32_t lds_size;
326 bool max_vert_out_per_gs_instance;
327 };
328
329 enum radv_shader_type {
330 RADV_SHADER_TYPE_DEFAULT = 0,
331 RADV_SHADER_TYPE_GS_COPY,
332 RADV_SHADER_TYPE_TRAP_HANDLER,
333 };
334
335 struct radv_shader_info {
336 uint64_t inline_push_constant_mask;
337 bool can_inline_all_push_constants;
338 bool loads_push_constants;
339 bool loads_dynamic_offsets;
340 uint32_t desc_set_used_mask;
341 bool uses_view_index;
342 bool uses_invocation_id;
343 bool uses_prim_id;
344 uint8_t wave_size;
345 uint8_t ballot_bit_size;
346 struct radv_userdata_locations user_sgprs_locs;
347 bool is_ngg;
348 bool is_ngg_passthrough;
349 bool has_ngg_culling;
350 bool has_ngg_early_prim_export;
351 bool has_prim_query;
352 bool has_xfb_query;
353 uint32_t num_tess_patches;
354 uint32_t esgs_itemsize; /* Only for VS or TES as ES */
355 struct radv_vs_output_info outinfo;
356 unsigned workgroup_size;
357 bool force_vrs_per_vertex;
358 gl_shader_stage stage;
359 gl_shader_stage next_stage;
360 enum radv_shader_type type;
361 uint32_t user_data_0;
362 bool inputs_linked;
363 bool outputs_linked;
364 bool has_epilog; /* Only for TCS or PS */
365 bool merged_shader_compiled_separately; /* GFX9+ */
366
367 struct {
368 uint8_t input_usage_mask[RADV_VERT_ATTRIB_MAX];
369 uint8_t output_usage_mask[VARYING_SLOT_VAR31 + 1];
370 bool needs_draw_id;
371 bool needs_instance_id;
372 bool as_es;
373 bool as_ls;
374 bool tcs_in_out_eq;
375 uint64_t tcs_temp_only_input_mask;
376 uint8_t num_linked_outputs;
377 bool needs_base_instance;
378 bool use_per_attribute_vb_descs;
379 uint32_t vb_desc_usage_mask;
380 uint32_t input_slot_usage_mask;
381 bool has_prolog;
382 bool dynamic_inputs;
383 bool dynamic_num_verts_per_prim;
384 uint32_t num_outputs; /* For NGG streamout only */
385 } vs;
386 struct {
387 uint8_t output_usage_mask[VARYING_SLOT_VAR31 + 1];
388 uint8_t num_stream_output_components[4];
389 uint8_t output_streams[VARYING_SLOT_VAR31 + 1];
390 uint8_t max_stream;
391 unsigned gsvs_vertex_size;
392 unsigned max_gsvs_emit_size;
393 unsigned vertices_in;
394 unsigned vertices_out;
395 unsigned input_prim;
396 unsigned output_prim;
397 unsigned invocations;
398 unsigned es_type; /* GFX9: VS or TES */
399 uint8_t num_linked_inputs;
400 bool has_pipeline_stat_query;
401 } gs;
402 struct {
403 uint8_t output_usage_mask[VARYING_SLOT_VAR31 + 1];
404 bool as_es;
405 enum tess_primitive_mode _primitive_mode;
406 enum gl_tess_spacing spacing;
407 bool ccw;
408 bool point_mode;
409 bool reads_tess_factors;
410 unsigned tcs_vertices_out;
411 uint8_t num_linked_inputs;
412 uint8_t num_linked_outputs;
413 uint32_t num_outputs; /* For NGG streamout only */
414 } tes;
415 struct {
416 bool uses_sample_shading;
417 bool needs_sample_positions;
418 bool needs_poly_line_smooth;
419 bool writes_memory;
420 bool writes_z;
421 bool writes_stencil;
422 bool writes_sample_mask;
423 bool writes_mrt0_alpha;
424 bool exports_mrtz_via_epilog;
425 bool has_pcoord;
426 bool prim_id_input;
427 bool layer_input;
428 bool viewport_index_input;
429 uint8_t num_input_clips_culls;
430 uint32_t input_mask;
431 uint32_t input_per_primitive_mask;
432 uint32_t flat_shaded_mask;
433 uint32_t explicit_shaded_mask;
434 uint32_t per_vertex_shaded_mask;
435 uint32_t float16_shaded_mask;
436 uint32_t num_interp;
437 uint32_t num_prim_interp;
438 bool can_discard;
439 bool early_fragment_test;
440 bool post_depth_coverage;
441 bool reads_sample_mask_in;
442 bool reads_front_face;
443 bool reads_sample_id;
444 bool reads_frag_shading_rate;
445 bool reads_barycentric_model;
446 bool reads_persp_sample;
447 bool reads_persp_center;
448 bool reads_persp_centroid;
449 bool reads_linear_sample;
450 bool reads_linear_center;
451 bool reads_linear_centroid;
452 bool reads_fully_covered;
453 uint8_t reads_frag_coord_mask;
454 uint8_t reads_sample_pos_mask;
455 uint8_t depth_layout;
456 bool allow_flat_shading;
457 bool pops; /* Uses Primitive Ordered Pixel Shading (fragment shader interlock) */
458 bool pops_is_per_sample;
459 bool mrt0_is_dual_src;
460 unsigned spi_ps_input;
461 unsigned colors_written;
462 unsigned spi_shader_col_format;
463 uint8_t color0_written;
464 bool load_provoking_vtx;
465 bool load_rasterization_prim;
466 bool force_sample_iter_shading_rate;
467 uint32_t db_shader_control; /* DB_SHADER_CONTROL without intrinsic rate overrides */
468 } ps;
469 struct {
470 bool uses_grid_size;
471 bool uses_block_id[3];
472 bool uses_thread_id[3];
473 bool uses_local_invocation_idx;
474 unsigned block_size[3];
475
476 bool is_rt_shader;
477 bool uses_dynamic_rt_callable_stack;
478 bool uses_rt;
479 bool uses_full_subgroups;
480 bool linear_taskmesh_dispatch;
481 bool has_query; /* Task shader only */
482
483 bool regalloc_hang_bug;
484 } cs;
485 struct {
486 uint64_t tes_inputs_read;
487 uint64_t tes_patch_inputs_read;
488 unsigned tcs_vertices_out;
489 uint32_t num_lds_blocks;
490 uint8_t num_linked_inputs;
491 uint8_t num_linked_outputs;
492 uint8_t num_linked_patch_outputs;
493 bool tes_reads_tess_factors : 1;
494 } tcs;
495 struct {
496 enum mesa_prim output_prim;
497 bool needs_ms_scratch_ring;
498 bool has_task; /* If mesh shader is used together with a task shader. */
499 bool has_query;
500 } ms;
501
502 struct radv_streamout_info so;
503
504 struct radv_legacy_gs_info gs_ring_info;
505 struct gfx10_ngg_info ngg_info;
506 };
507
508 struct radv_vs_input_state {
509 uint32_t attribute_mask;
510
511 uint32_t instance_rate_inputs;
512 uint32_t nontrivial_divisors;
513 uint32_t zero_divisors;
514 uint32_t post_shuffle;
515 /* Having two separate fields instead of a single uint64_t makes it easier to remove attributes
516 * using bitwise arithmetic.
517 */
518 uint32_t alpha_adjust_lo;
519 uint32_t alpha_adjust_hi;
520 uint32_t nontrivial_formats;
521
522 uint8_t bindings[MAX_VERTEX_ATTRIBS];
523 uint32_t divisors[MAX_VERTEX_ATTRIBS];
524 uint32_t offsets[MAX_VERTEX_ATTRIBS];
525 uint8_t formats[MAX_VERTEX_ATTRIBS];
526 uint8_t format_align_req_minus_1[MAX_VERTEX_ATTRIBS];
527 uint8_t format_sizes[MAX_VERTEX_ATTRIBS];
528
529 bool bindings_match_attrib;
530 };
531
532 struct radv_vs_prolog_key {
533 /* All the fields are pre-masked with BITFIELD_MASK(num_attributes).
534 * Some of the fields are pre-masked by other conditions. See lookup_vs_prolog.
535 */
536 uint32_t instance_rate_inputs;
537 uint32_t nontrivial_divisors;
538 uint32_t zero_divisors;
539 uint32_t post_shuffle;
540 /* Having two separate fields instead of a single uint64_t makes it easier to remove attributes
541 * using bitwise arithmetic.
542 */
543 uint32_t alpha_adjust_lo;
544 uint32_t alpha_adjust_hi;
545 uint8_t formats[MAX_VERTEX_ATTRIBS];
546 unsigned num_attributes;
547 uint32_t misaligned_mask;
548 bool as_ls;
549 bool is_ngg;
550 bool wave32;
551 gl_shader_stage next_stage;
552 };
553
554 struct radv_tcs_epilog_key {
555 enum tess_primitive_mode primitive_mode;
556 bool tes_reads_tessfactors;
557 bool tcs_out_patch_fits_subgroup;
558 };
559
560 enum radv_shader_binary_type { RADV_BINARY_TYPE_LEGACY, RADV_BINARY_TYPE_RTLD };
561
562 struct radv_shader_binary {
563 uint32_t type; /* enum radv_shader_binary_type */
564
565 struct ac_shader_config config;
566 struct radv_shader_info info;
567
568 /* Self-referential size so we avoid consistency issues. */
569 uint32_t total_size;
570 };
571
572 struct radv_shader_binary_legacy {
573 struct radv_shader_binary base;
574 uint32_t code_size;
575 uint32_t exec_size;
576 uint32_t ir_size;
577 uint32_t disasm_size;
578 uint32_t stats_size;
579 uint32_t padding;
580
581 /* data has size of stats_size + code_size + ir_size + disasm_size + 2,
582 * where the +2 is for 0 of the ir strings. */
583 uint8_t data[0];
584 };
585 static_assert(sizeof(struct radv_shader_binary_legacy) == offsetof(struct radv_shader_binary_legacy, data),
586 "Unexpected padding");
587
588 struct radv_shader_binary_rtld {
589 struct radv_shader_binary base;
590 unsigned elf_size;
591 unsigned llvm_ir_size;
592 uint8_t data[0];
593 };
594
595 struct radv_shader_part_binary {
596 struct {
597 uint32_t spi_shader_col_format;
598 uint32_t spi_shader_z_format;
599 } info;
600
601 uint8_t num_sgprs;
602 uint8_t num_vgprs;
603 unsigned code_size;
604 unsigned disasm_size;
605
606 /* Self-referential size so we avoid consistency issues. */
607 uint32_t total_size;
608
609 uint8_t data[0];
610 };
611
612 enum radv_shader_arena_type { RADV_SHADER_ARENA_DEFAULT, RADV_SHADER_ARENA_REPLAYABLE, RADV_SHADER_ARENA_REPLAYED };
613
614 struct radv_shader_arena {
615 struct list_head list;
616 struct list_head entries;
617 uint32_t size;
618 struct radeon_winsys_bo *bo;
619 char *ptr;
620 enum radv_shader_arena_type type;
621 };
622
623 union radv_shader_arena_block {
624 struct list_head pool;
625 struct {
626 /* List of blocks in the arena, sorted by address. */
627 struct list_head list;
628 /* For holes, a list_head for the free-list. For allocations, freelist.prev=NULL and
629 * freelist.next is a pointer associated with the allocation.
630 */
631 struct list_head freelist;
632 struct radv_shader_arena *arena;
633 uint32_t offset;
634 uint32_t size;
635 };
636 };
637
638 struct radv_shader_free_list {
639 uint8_t size_mask;
640 struct list_head free_lists[RADV_SHADER_ALLOC_NUM_FREE_LISTS];
641 };
642
643 struct radv_serialized_shader_arena_block {
644 uint32_t offset;
645 uint32_t size;
646 uint64_t arena_va;
647 uint32_t arena_size;
648 };
649
650 struct radv_shader {
651 struct vk_pipeline_cache_object base;
652
653 simple_mtx_t replay_mtx;
654 bool has_replay_alloc;
655
656 struct radeon_winsys_bo *bo;
657 union radv_shader_arena_block *alloc;
658 uint64_t va;
659
660 uint64_t upload_seq;
661
662 struct ac_shader_config config;
663 uint32_t code_size;
664 uint32_t exec_size;
665 struct radv_shader_info info;
666 uint32_t max_waves;
667
668 blake3_hash hash;
669 void *code;
670
671 /* debug only */
672 char *spirv;
673 uint32_t spirv_size;
674 char *nir_string;
675 char *disasm_string;
676 char *ir_string;
677 uint32_t *statistics;
678 };
679
680 struct radv_shader_part {
681 uint32_t ref_count;
682
683 union {
684 struct radv_vs_prolog_key vs;
685 struct radv_ps_epilog_key ps;
686 struct radv_tcs_epilog_key tcs;
687 } key;
688
689 uint64_t va;
690
691 struct radeon_winsys_bo *bo;
692 union radv_shader_arena_block *alloc;
693 uint32_t code_size;
694 uint32_t rsrc1;
695 bool nontrivial_divisors;
696 uint32_t spi_shader_col_format;
697 uint32_t spi_shader_z_format;
698 uint64_t upload_seq;
699
700 /* debug only */
701 char *disasm_string;
702 };
703
704 struct radv_shader_part_cache_ops {
705 uint32_t (*hash)(const void *key);
706 bool (*equals)(const void *a, const void *b);
707 struct radv_shader_part *(*create)(struct radv_device *device, const void *key);
708 };
709
710 struct radv_shader_part_cache {
711 simple_mtx_t lock;
712 struct radv_shader_part_cache_ops *ops;
713 struct set entries;
714 };
715
716 struct radv_shader_dma_submission {
717 struct list_head list;
718
719 struct radeon_cmdbuf *cs;
720 struct radeon_winsys_bo *bo;
721 uint64_t bo_size;
722 char *ptr;
723
724 /* The semaphore value to wait for before reusing this submission. */
725 uint64_t seq;
726 };
727
728 struct radv_shader_object {
729 struct vk_object_base base;
730
731 gl_shader_stage stage;
732
733 VkShaderCodeTypeEXT code_type;
734
735 /* Main shader */
736 struct radv_shader *shader;
737 struct radv_shader_binary *binary;
738
739 /* Shader variants */
740 /* VS before TCS */
741 struct {
742 struct radv_shader *shader;
743 struct radv_shader_binary *binary;
744 } as_ls;
745
746 /* VS/TES before GS */
747 struct {
748 struct radv_shader *shader;
749 struct radv_shader_binary *binary;
750 } as_es;
751
752 /* GS copy shader */
753 struct {
754 struct radv_shader *copy_shader;
755 struct radv_shader_binary *copy_binary;
756 } gs;
757
758 uint32_t push_constant_size;
759 uint32_t dynamic_offset_count;
760 };
761
762 struct radv_pipeline_layout;
763 struct radv_shader_stage;
764
765 void radv_optimize_nir(struct nir_shader *shader, bool optimize_conservatively);
766 void radv_optimize_nir_algebraic(nir_shader *shader, bool opt_offsets);
767
768 void radv_postprocess_nir(struct radv_device *device, const struct radv_graphics_state_key *gfx_state,
769 struct radv_shader_stage *stage);
770
771 bool radv_shader_should_clear_lds(const struct radv_device *device, const nir_shader *shader);
772
773 void radv_nir_lower_rt_io(nir_shader *shader, bool monolithic, uint32_t payload_offset);
774
775 void radv_nir_lower_rt_abi(nir_shader *shader, const VkRayTracingPipelineCreateInfoKHR *pCreateInfo,
776 const struct radv_shader_args *args, const struct radv_shader_info *info,
777 uint32_t *stack_size, bool resume_shader, struct radv_device *device,
778 struct radv_ray_tracing_pipeline *pipeline, bool monolithic);
779
780 struct radv_shader_stage;
781
782 nir_shader *radv_shader_spirv_to_nir(struct radv_device *device, const struct radv_shader_stage *stage,
783 const struct radv_spirv_to_nir_options *options, bool is_internal);
784
785 void radv_init_shader_arenas(struct radv_device *device);
786 void radv_destroy_shader_arenas(struct radv_device *device);
787 VkResult radv_init_shader_upload_queue(struct radv_device *device);
788 void radv_destroy_shader_upload_queue(struct radv_device *device);
789
790 struct radv_shader_args;
791
792 struct radv_shader *radv_shader_create(struct radv_device *device, struct vk_pipeline_cache *cache,
793 const struct radv_shader_binary *binary, bool skip_cache);
794
795 VkResult radv_shader_create_uncached(struct radv_device *device, const struct radv_shader_binary *binary,
796 bool replayable, struct radv_serialized_shader_arena_block *replay_block,
797 struct radv_shader **out_shader);
798
799 struct radv_shader_binary *radv_shader_nir_to_asm(struct radv_device *device, struct radv_shader_stage *pl_stage,
800 struct nir_shader *const *shaders, int shader_count,
801 const struct radv_graphics_state_key *gfx_state,
802 bool keep_shader_info, bool keep_statistic_info);
803
804 void radv_shader_generate_debug_info(struct radv_device *device, bool dump_shader, bool keep_shader_info,
805 struct radv_shader_binary *binary, struct radv_shader *shader,
806 struct nir_shader *const *shaders, int shader_count,
807 struct radv_shader_info *info);
808
809 VkResult radv_shader_wait_for_upload(struct radv_device *device, uint64_t seq);
810
811 struct radv_shader_dma_submission *radv_shader_dma_pop_submission(struct radv_device *device);
812
813 void radv_shader_dma_push_submission(struct radv_device *device, struct radv_shader_dma_submission *submission,
814 uint64_t seq);
815
816 struct radv_shader_dma_submission *
817 radv_shader_dma_get_submission(struct radv_device *device, struct radeon_winsys_bo *bo, uint64_t va, uint64_t size);
818
819 bool radv_shader_dma_submit(struct radv_device *device, struct radv_shader_dma_submission *submission,
820 uint64_t *upload_seq_out);
821
822 union radv_shader_arena_block *radv_alloc_shader_memory(struct radv_device *device, uint32_t size, bool replayable,
823 void *ptr);
824
825 union radv_shader_arena_block *radv_replay_shader_arena_block(struct radv_device *device,
826 const struct radv_serialized_shader_arena_block *src,
827 void *ptr);
828
829 struct radv_serialized_shader_arena_block radv_serialize_shader_arena_block(union radv_shader_arena_block *block);
830
831 void radv_free_shader_memory(struct radv_device *device, union radv_shader_arena_block *alloc);
832
833 struct radv_shader *radv_create_trap_handler_shader(struct radv_device *device);
834
835 struct radv_shader *radv_create_rt_prolog(struct radv_device *device);
836
837 struct radv_shader_part *radv_shader_part_create(struct radv_device *device, struct radv_shader_part_binary *binary,
838 unsigned wave_size);
839
840 struct radv_shader_part *radv_create_vs_prolog(struct radv_device *device, const struct radv_vs_prolog_key *key);
841
842 struct radv_shader_part *radv_create_ps_epilog(struct radv_device *device, const struct radv_ps_epilog_key *key,
843 struct radv_shader_part_binary **binary_out);
844
845 struct radv_shader_part *radv_create_tcs_epilog(struct radv_device *device, const struct radv_tcs_epilog_key *key);
846
847 void radv_shader_part_destroy(struct radv_device *device, struct radv_shader_part *shader_part);
848
849 bool radv_shader_part_cache_init(struct radv_shader_part_cache *cache, struct radv_shader_part_cache_ops *ops);
850 void radv_shader_part_cache_finish(struct radv_device *device, struct radv_shader_part_cache *cache);
851 struct radv_shader_part *radv_shader_part_cache_get(struct radv_device *device, struct radv_shader_part_cache *cache,
852 struct set *local_entries, const void *key);
853
854 uint64_t radv_shader_get_va(const struct radv_shader *shader);
855 struct radv_shader *radv_find_shader(struct radv_device *device, uint64_t pc);
856
857 unsigned radv_get_max_waves(const struct radv_device *device, const struct ac_shader_config *conf,
858 const struct radv_shader_info *info);
859
860 unsigned radv_get_max_scratch_waves(const struct radv_device *device, struct radv_shader *shader);
861
862 const char *radv_get_shader_name(const struct radv_shader_info *info, gl_shader_stage stage);
863
864 unsigned radv_compute_spi_ps_input(const struct radv_graphics_state_key *gfx_state,
865 const struct radv_shader_info *info);
866
867 bool radv_can_dump_shader(struct radv_device *device, nir_shader *nir, bool meta_shader);
868
869 bool radv_can_dump_shader_stats(struct radv_device *device, nir_shader *nir);
870
871 VkResult radv_dump_shader_stats(struct radv_device *device, struct radv_pipeline *pipeline, struct radv_shader *shader,
872 gl_shader_stage stage, FILE *output);
873
874 /* Returns true on success and false on failure */
875 bool radv_shader_reupload(struct radv_device *device, struct radv_shader *shader);
876
877 enum ac_hw_stage radv_select_hw_stage(const struct radv_shader_info *const info, const enum amd_gfx_level gfx_level);
878
879 extern const struct vk_pipeline_cache_object_ops radv_shader_ops;
880
881 static inline struct radv_shader *
radv_shader_ref(struct radv_shader * shader)882 radv_shader_ref(struct radv_shader *shader)
883 {
884 vk_pipeline_cache_object_ref(&shader->base);
885 return shader;
886 }
887
888 static inline void
radv_shader_unref(struct radv_device * device,struct radv_shader * shader)889 radv_shader_unref(struct radv_device *device, struct radv_shader *shader)
890 {
891 vk_pipeline_cache_object_unref((struct vk_device *)device, &shader->base);
892 }
893
894 static inline struct radv_shader_part *
radv_shader_part_ref(struct radv_shader_part * shader_part)895 radv_shader_part_ref(struct radv_shader_part *shader_part)
896 {
897 assert(shader_part && shader_part->ref_count >= 1);
898 p_atomic_inc(&shader_part->ref_count);
899 return shader_part;
900 }
901
902 static inline void
radv_shader_part_unref(struct radv_device * device,struct radv_shader_part * shader_part)903 radv_shader_part_unref(struct radv_device *device, struct radv_shader_part *shader_part)
904 {
905 assert(shader_part && shader_part->ref_count >= 1);
906 if (p_atomic_dec_zero(&shader_part->ref_count))
907 radv_shader_part_destroy(device, shader_part);
908 }
909
910 static inline struct radv_shader_part *
radv_shader_part_from_cache_entry(const void * key)911 radv_shader_part_from_cache_entry(const void *key)
912 {
913 return container_of(key, struct radv_shader_part, key);
914 }
915
916 static inline unsigned
get_tcs_input_vertex_stride(unsigned tcs_num_inputs)917 get_tcs_input_vertex_stride(unsigned tcs_num_inputs)
918 {
919 unsigned stride = tcs_num_inputs * 16;
920
921 /* Add 1 dword to reduce LDS bank conflicts. */
922 if (stride)
923 stride += 4;
924
925 return stride;
926 }
927
928 static inline unsigned
calculate_tess_lds_size(enum amd_gfx_level gfx_level,unsigned tcs_num_input_vertices,unsigned tcs_num_output_vertices,unsigned tcs_num_inputs,unsigned tcs_num_patches,unsigned tcs_num_outputs,unsigned tcs_num_patch_outputs)929 calculate_tess_lds_size(enum amd_gfx_level gfx_level, unsigned tcs_num_input_vertices, unsigned tcs_num_output_vertices,
930 unsigned tcs_num_inputs, unsigned tcs_num_patches, unsigned tcs_num_outputs,
931 unsigned tcs_num_patch_outputs)
932 {
933 unsigned input_vertex_size = get_tcs_input_vertex_stride(tcs_num_inputs);
934 unsigned output_vertex_size = tcs_num_outputs * 16;
935
936 unsigned input_patch_size = tcs_num_input_vertices * input_vertex_size;
937
938 unsigned pervertex_output_patch_size = tcs_num_output_vertices * output_vertex_size;
939 unsigned output_patch_size = pervertex_output_patch_size + tcs_num_patch_outputs * 16;
940
941 unsigned output_patch0_offset = input_patch_size * tcs_num_patches;
942
943 unsigned lds_size = output_patch0_offset + output_patch_size * tcs_num_patches;
944
945 if (gfx_level >= GFX7) {
946 assert(lds_size <= 65536);
947 lds_size = align(lds_size, 512) / 512;
948 } else {
949 assert(lds_size <= 32768);
950 lds_size = align(lds_size, 256) / 256;
951 }
952
953 return lds_size;
954 }
955
956 static inline unsigned
get_tcs_num_patches(unsigned tcs_num_input_vertices,unsigned tcs_num_output_vertices,unsigned tcs_num_inputs,unsigned tcs_num_outputs,unsigned tcs_num_patch_outputs,unsigned tess_offchip_block_dw_size,enum amd_gfx_level gfx_level,enum radeon_family family)957 get_tcs_num_patches(unsigned tcs_num_input_vertices, unsigned tcs_num_output_vertices, unsigned tcs_num_inputs,
958 unsigned tcs_num_outputs, unsigned tcs_num_patch_outputs, unsigned tess_offchip_block_dw_size,
959 enum amd_gfx_level gfx_level, enum radeon_family family)
960 {
961 uint32_t input_vertex_size = get_tcs_input_vertex_stride(tcs_num_inputs);
962 uint32_t input_patch_size = tcs_num_input_vertices * input_vertex_size;
963 uint32_t output_vertex_size = tcs_num_outputs * 16;
964 uint32_t pervertex_output_patch_size = tcs_num_output_vertices * output_vertex_size;
965 uint32_t output_patch_size = pervertex_output_patch_size + tcs_num_patch_outputs * 16;
966
967 /* Ensure that we only need one wave per SIMD so we don't need to check
968 * resource usage. Also ensures that the number of tcs in and out
969 * vertices per threadgroup are at most 256.
970 */
971 unsigned num_patches = 64 / MAX2(tcs_num_input_vertices, tcs_num_output_vertices) * 4;
972 /* Make sure that the data fits in LDS. This assumes the shaders only
973 * use LDS for the inputs and outputs.
974 */
975 unsigned hardware_lds_size = 32768;
976
977 /* Looks like STONEY hangs if we use more than 32 KiB LDS in a single
978 * threadgroup, even though there is more than 32 KiB LDS.
979 *
980 * Test: dEQP-VK.tessellation.shader_input_output.barrier
981 */
982 if (gfx_level >= GFX7 && family != CHIP_STONEY)
983 hardware_lds_size = 65536;
984
985 if (input_patch_size + output_patch_size)
986 num_patches = MIN2(num_patches, hardware_lds_size / (input_patch_size + output_patch_size));
987 /* Make sure the output data fits in the offchip buffer */
988 if (output_patch_size)
989 num_patches = MIN2(num_patches, (tess_offchip_block_dw_size * 4) / output_patch_size);
990 /* Not necessary for correctness, but improves performance. The
991 * specific value is taken from the proprietary driver.
992 */
993 num_patches = MIN2(num_patches, 40);
994
995 /* GFX6 bug workaround - limit LS-HS threadgroups to only one wave. */
996 if (gfx_level == GFX6) {
997 unsigned one_wave = 64 / MAX2(tcs_num_input_vertices, tcs_num_output_vertices);
998 num_patches = MIN2(num_patches, one_wave);
999 }
1000 return num_patches;
1001 }
1002
1003 void radv_lower_ngg(struct radv_device *device, struct radv_shader_stage *ngg_stage,
1004 const struct radv_graphics_state_key *gfx_state);
1005
1006 bool radv_consider_culling(const struct radv_physical_device *pdevice, struct nir_shader *nir, uint64_t ps_inputs_read,
1007 unsigned num_vertices_per_primitive, const struct radv_shader_info *info);
1008
1009 void radv_get_nir_options(struct radv_physical_device *device);
1010
1011 nir_shader *radv_build_traversal_shader(struct radv_device *device, struct radv_ray_tracing_pipeline *pipeline,
1012 const VkRayTracingPipelineCreateInfoKHR *pCreateInfo);
1013
1014 enum radv_rt_priority {
1015 radv_rt_priority_raygen = 0,
1016 radv_rt_priority_traversal = 1,
1017 radv_rt_priority_hit_miss = 2,
1018 radv_rt_priority_callable = 3,
1019 radv_rt_priority_mask = 0x3,
1020 };
1021
1022 static inline enum radv_rt_priority
radv_get_rt_priority(gl_shader_stage stage)1023 radv_get_rt_priority(gl_shader_stage stage)
1024 {
1025 switch (stage) {
1026 case MESA_SHADER_RAYGEN:
1027 return radv_rt_priority_raygen;
1028 case MESA_SHADER_INTERSECTION:
1029 case MESA_SHADER_ANY_HIT:
1030 return radv_rt_priority_traversal;
1031 case MESA_SHADER_CLOSEST_HIT:
1032 case MESA_SHADER_MISS:
1033 return radv_rt_priority_hit_miss;
1034 case MESA_SHADER_CALLABLE:
1035 return radv_rt_priority_callable;
1036 default:
1037 unreachable("Unimplemented RT shader stage.");
1038 }
1039 }
1040
1041 struct radv_shader_layout;
1042 enum radv_pipeline_type;
1043
1044 void radv_nir_shader_info_pass(struct radv_device *device, const struct nir_shader *nir,
1045 const struct radv_shader_layout *layout, const struct radv_shader_stage_key *stage_key,
1046 const struct radv_graphics_state_key *gfx_state,
1047 const enum radv_pipeline_type pipeline_type, bool consider_force_vrs,
1048 struct radv_shader_info *info);
1049
1050 void radv_nir_shader_info_init(gl_shader_stage stage, gl_shader_stage next_stage, struct radv_shader_info *info);
1051
1052 void radv_nir_shader_info_link(struct radv_device *device, const struct radv_graphics_state_key *gfx_state,
1053 struct radv_shader_stage *stages);
1054
1055 void radv_shader_combine_cfg_vs_tcs(const struct radv_shader *vs, const struct radv_shader *tcs, uint32_t *rsrc1_out,
1056 uint32_t *rsrc2_out);
1057
1058 void radv_shader_combine_cfg_vs_gs(const struct radv_shader *vs, const struct radv_shader *gs, uint32_t *rsrc1_out,
1059 uint32_t *rsrc2_out);
1060
1061 void radv_shader_combine_cfg_tes_gs(const struct radv_shader *tes, const struct radv_shader *gs, uint32_t *rsrc1_out,
1062 uint32_t *rsrc2_out);
1063
1064 #endif
1065