1 /* 2 * Copyright © 2020 Valve Corporation 3 * 4 * Permission is hereby granted, free of charge, to any person obtaining a 5 * copy of this software and associated documentation files (the "Software"), 6 * to deal in the Software without restriction, including without limitation 7 * the rights to use, copy, modify, merge, publish, distribute, sublicense, 8 * and/or sell copies of the Software, and to permit persons to whom the 9 * Software is furnished to do so, subject to the following conditions: 10 * 11 * The above copyright notice and this permission notice (including the next 12 * paragraph) shall be included in all copies or substantial portions of the 13 * Software. 14 * 15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING 20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS 21 * IN THE SOFTWARE. 22 * 23 */ 24 25 /* 26 * Replaces make availability/visible semantics on barriers with 27 * ACCESS_COHERENT on memory loads/stores 28 */ 29 30 #include "nir/nir.h" 31 #include "shader_enums.h" 32 33 static bool get_intrinsic_info(nir_intrinsic_instr * intrin,nir_variable_mode * modes,bool * reads,bool * writes)34 get_intrinsic_info(nir_intrinsic_instr *intrin, nir_variable_mode *modes, 35 bool *reads, bool *writes) 36 { 37 switch (intrin->intrinsic) { 38 case nir_intrinsic_image_deref_load: 39 case nir_intrinsic_image_deref_sparse_load: 40 *modes = nir_src_as_deref(intrin->src[0])->modes; 41 *reads = true; 42 break; 43 case nir_intrinsic_image_deref_store: 44 *modes = nir_src_as_deref(intrin->src[0])->modes; 45 *writes = true; 46 break; 47 case nir_intrinsic_image_deref_atomic_add: 48 case nir_intrinsic_image_deref_atomic_fadd: 49 case nir_intrinsic_image_deref_atomic_umin: 50 case nir_intrinsic_image_deref_atomic_imin: 51 case nir_intrinsic_image_deref_atomic_umax: 52 case nir_intrinsic_image_deref_atomic_imax: 53 case nir_intrinsic_image_deref_atomic_fmin: 54 case nir_intrinsic_image_deref_atomic_fmax: 55 case nir_intrinsic_image_deref_atomic_and: 56 case nir_intrinsic_image_deref_atomic_or: 57 case nir_intrinsic_image_deref_atomic_xor: 58 case nir_intrinsic_image_deref_atomic_exchange: 59 case nir_intrinsic_image_deref_atomic_comp_swap: 60 *modes = nir_src_as_deref(intrin->src[0])->modes; 61 *reads = true; 62 *writes = true; 63 break; 64 case nir_intrinsic_load_ssbo: 65 *modes = nir_var_mem_ssbo; 66 *reads = true; 67 break; 68 case nir_intrinsic_store_ssbo: 69 *modes = nir_var_mem_ssbo; 70 *writes = true; 71 break; 72 case nir_intrinsic_ssbo_atomic_add: 73 case nir_intrinsic_ssbo_atomic_imin: 74 case nir_intrinsic_ssbo_atomic_umin: 75 case nir_intrinsic_ssbo_atomic_imax: 76 case nir_intrinsic_ssbo_atomic_umax: 77 case nir_intrinsic_ssbo_atomic_and: 78 case nir_intrinsic_ssbo_atomic_or: 79 case nir_intrinsic_ssbo_atomic_xor: 80 case nir_intrinsic_ssbo_atomic_exchange: 81 case nir_intrinsic_ssbo_atomic_comp_swap: 82 case nir_intrinsic_ssbo_atomic_fadd: 83 case nir_intrinsic_ssbo_atomic_fcomp_swap: 84 case nir_intrinsic_ssbo_atomic_fmax: 85 case nir_intrinsic_ssbo_atomic_fmin: 86 *modes = nir_var_mem_ssbo; 87 *reads = true; 88 *writes = true; 89 break; 90 case nir_intrinsic_load_global: 91 *modes = nir_var_mem_global; 92 *reads = true; 93 break; 94 case nir_intrinsic_store_global: 95 *modes = nir_var_mem_global; 96 *writes = true; 97 break; 98 case nir_intrinsic_global_atomic_add: 99 case nir_intrinsic_global_atomic_imin: 100 case nir_intrinsic_global_atomic_umin: 101 case nir_intrinsic_global_atomic_imax: 102 case nir_intrinsic_global_atomic_umax: 103 case nir_intrinsic_global_atomic_and: 104 case nir_intrinsic_global_atomic_or: 105 case nir_intrinsic_global_atomic_xor: 106 case nir_intrinsic_global_atomic_exchange: 107 case nir_intrinsic_global_atomic_comp_swap: 108 case nir_intrinsic_global_atomic_fadd: 109 case nir_intrinsic_global_atomic_fcomp_swap: 110 case nir_intrinsic_global_atomic_fmax: 111 case nir_intrinsic_global_atomic_fmin: 112 *modes = nir_var_mem_global; 113 *reads = true; 114 *writes = true; 115 break; 116 case nir_intrinsic_load_deref: 117 *modes = nir_src_as_deref(intrin->src[0])->modes; 118 *reads = true; 119 break; 120 case nir_intrinsic_store_deref: 121 *modes = nir_src_as_deref(intrin->src[0])->modes; 122 *writes = true; 123 break; 124 case nir_intrinsic_deref_atomic_add: 125 case nir_intrinsic_deref_atomic_imin: 126 case nir_intrinsic_deref_atomic_umin: 127 case nir_intrinsic_deref_atomic_imax: 128 case nir_intrinsic_deref_atomic_umax: 129 case nir_intrinsic_deref_atomic_and: 130 case nir_intrinsic_deref_atomic_or: 131 case nir_intrinsic_deref_atomic_xor: 132 case nir_intrinsic_deref_atomic_exchange: 133 case nir_intrinsic_deref_atomic_comp_swap: 134 case nir_intrinsic_deref_atomic_fadd: 135 case nir_intrinsic_deref_atomic_fmin: 136 case nir_intrinsic_deref_atomic_fmax: 137 case nir_intrinsic_deref_atomic_fcomp_swap: 138 *modes = nir_src_as_deref(intrin->src[0])->modes; 139 *reads = true; 140 *writes = true; 141 break; 142 default: 143 return false; 144 } 145 return true; 146 } 147 148 static bool visit_instr(nir_instr * instr,uint32_t * cur_modes,unsigned vis_avail_sem)149 visit_instr(nir_instr *instr, uint32_t *cur_modes, unsigned vis_avail_sem) 150 { 151 if (instr->type != nir_instr_type_intrinsic) 152 return false; 153 nir_intrinsic_instr *intrin = nir_instr_as_intrinsic(instr); 154 155 if (intrin->intrinsic == nir_intrinsic_scoped_barrier && 156 (nir_intrinsic_memory_semantics(intrin) & vis_avail_sem)) { 157 *cur_modes |= nir_intrinsic_memory_modes(intrin); 158 159 unsigned semantics = nir_intrinsic_memory_semantics(intrin); 160 nir_intrinsic_set_memory_semantics( 161 intrin, semantics & ~vis_avail_sem); 162 return true; 163 } 164 165 if (!*cur_modes) 166 return false; /* early exit */ 167 168 nir_variable_mode modes; 169 bool reads = false, writes = false; 170 if (!get_intrinsic_info(intrin, &modes, &reads, &writes)) 171 return false; 172 173 if (!reads && vis_avail_sem == NIR_MEMORY_MAKE_VISIBLE) 174 return false; 175 if (!writes && vis_avail_sem == NIR_MEMORY_MAKE_AVAILABLE) 176 return false; 177 178 if (!nir_intrinsic_has_access(intrin)) 179 return false; 180 181 unsigned access = nir_intrinsic_access(intrin); 182 183 if (access & (ACCESS_NON_READABLE | ACCESS_NON_WRITEABLE | ACCESS_CAN_REORDER | ACCESS_COHERENT)) 184 return false; 185 186 if (*cur_modes & modes) { 187 nir_intrinsic_set_access(intrin, access | ACCESS_COHERENT); 188 return true; 189 } 190 191 return false; 192 } 193 194 static bool lower_make_visible(nir_cf_node * cf_node,uint32_t * cur_modes)195 lower_make_visible(nir_cf_node *cf_node, uint32_t *cur_modes) 196 { 197 bool progress = false; 198 switch (cf_node->type) { 199 case nir_cf_node_block: { 200 nir_block *block = nir_cf_node_as_block(cf_node); 201 nir_foreach_instr(instr, block) 202 progress |= visit_instr(instr, cur_modes, NIR_MEMORY_MAKE_VISIBLE); 203 break; 204 } 205 case nir_cf_node_if: { 206 nir_if *nif = nir_cf_node_as_if(cf_node); 207 uint32_t cur_modes_then = *cur_modes; 208 uint32_t cur_modes_else = *cur_modes; 209 foreach_list_typed(nir_cf_node, if_node, node, &nif->then_list) 210 progress |= lower_make_visible(if_node, &cur_modes_then); 211 foreach_list_typed(nir_cf_node, if_node, node, &nif->else_list) 212 progress |= lower_make_visible(if_node, &cur_modes_else); 213 *cur_modes |= cur_modes_then | cur_modes_else; 214 break; 215 } 216 case nir_cf_node_loop: { 217 nir_loop *loop = nir_cf_node_as_loop(cf_node); 218 bool loop_progress; 219 do { 220 loop_progress = false; 221 foreach_list_typed(nir_cf_node, loop_node, node, &loop->body) 222 loop_progress |= lower_make_visible(loop_node, cur_modes); 223 progress |= loop_progress; 224 } while (loop_progress); 225 break; 226 } 227 case nir_cf_node_function: 228 unreachable("Invalid cf type"); 229 } 230 return progress; 231 } 232 233 static bool lower_make_available(nir_cf_node * cf_node,uint32_t * cur_modes)234 lower_make_available(nir_cf_node *cf_node, uint32_t *cur_modes) 235 { 236 bool progress = false; 237 switch (cf_node->type) { 238 case nir_cf_node_block: { 239 nir_block *block = nir_cf_node_as_block(cf_node); 240 nir_foreach_instr_reverse(instr, block) 241 progress |= visit_instr(instr, cur_modes, NIR_MEMORY_MAKE_AVAILABLE); 242 break; 243 } 244 case nir_cf_node_if: { 245 nir_if *nif = nir_cf_node_as_if(cf_node); 246 uint32_t cur_modes_then = *cur_modes; 247 uint32_t cur_modes_else = *cur_modes; 248 foreach_list_typed_reverse(nir_cf_node, if_node, node, &nif->then_list) 249 progress |= lower_make_available(if_node, &cur_modes_then); 250 foreach_list_typed_reverse(nir_cf_node, if_node, node, &nif->else_list) 251 progress |= lower_make_available(if_node, &cur_modes_else); 252 *cur_modes |= cur_modes_then | cur_modes_else; 253 break; 254 } 255 case nir_cf_node_loop: { 256 nir_loop *loop = nir_cf_node_as_loop(cf_node); 257 bool loop_progress; 258 do { 259 loop_progress = false; 260 foreach_list_typed_reverse(nir_cf_node, loop_node, node, &loop->body) 261 loop_progress |= lower_make_available(loop_node, cur_modes); 262 progress |= loop_progress; 263 } while (loop_progress); 264 break; 265 } 266 case nir_cf_node_function: 267 unreachable("Invalid cf type"); 268 } 269 return progress; 270 } 271 272 bool nir_lower_memory_model(nir_shader * shader)273 nir_lower_memory_model(nir_shader *shader) 274 { 275 bool progress = false; 276 277 struct exec_list *cf_list = &nir_shader_get_entrypoint(shader)->body; 278 279 uint32_t modes = 0; 280 foreach_list_typed(nir_cf_node, cf_node, node, cf_list) 281 progress |= lower_make_visible(cf_node, &modes); 282 283 modes = 0; 284 foreach_list_typed_reverse(nir_cf_node, cf_node, node, cf_list) 285 progress |= lower_make_available(cf_node, &modes); 286 287 return progress; 288 } 289