1 /*
2 * Copyright © 2015 Red Hat
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
21 * IN THE SOFTWARE.
22 */
23
24 #include "nir.h"
25 #include "nir_control_flow.h"
26
27 /* Secret Decoder Ring:
28 * clone_foo():
29 * Allocate and clone a foo.
30 * __clone_foo():
31 * Clone body of foo (ie. parent class, embedded struct, etc)
32 */
33
34 typedef struct {
35 /* True if we are cloning an entire shader. */
36 bool global_clone;
37
38 /* If true allows the clone operation to fall back to the original pointer
39 * if no clone pointer is found in the remap table. This allows us to
40 * clone a loop body without having to add srcs from outside the loop to
41 * the remap table. This is useful for loop unrolling.
42 */
43 bool allow_remap_fallback;
44
45 /* maps orig ptr -> cloned ptr: */
46 struct hash_table *remap_table;
47
48 /* List of phi sources. */
49 struct list_head phi_srcs;
50
51 /* new shader object, used as memctx for just about everything else: */
52 nir_shader *ns;
53 } clone_state;
54
55 static void
init_clone_state(clone_state * state,struct hash_table * remap_table,bool global,bool allow_remap_fallback)56 init_clone_state(clone_state *state, struct hash_table *remap_table,
57 bool global, bool allow_remap_fallback)
58 {
59 state->global_clone = global;
60 state->allow_remap_fallback = allow_remap_fallback;
61
62 if (remap_table) {
63 state->remap_table = remap_table;
64 } else {
65 state->remap_table = _mesa_pointer_hash_table_create(NULL);
66 }
67
68 list_inithead(&state->phi_srcs);
69 }
70
71 static void
free_clone_state(clone_state * state)72 free_clone_state(clone_state *state)
73 {
74 _mesa_hash_table_destroy(state->remap_table, NULL);
75 }
76
77 static inline void *
_lookup_ptr(clone_state * state,const void * ptr,bool global)78 _lookup_ptr(clone_state *state, const void *ptr, bool global)
79 {
80 struct hash_entry *entry;
81
82 if (!ptr)
83 return NULL;
84
85 if (!state->global_clone && global)
86 return (void *)ptr;
87
88 if (unlikely(!state->remap_table)) {
89 assert(state->allow_remap_fallback);
90 return (void *)ptr;
91 }
92
93 entry = _mesa_hash_table_search(state->remap_table, ptr);
94 if (!entry) {
95 assert(state->allow_remap_fallback);
96 return (void *)ptr;
97 }
98
99 return entry->data;
100 }
101
102 static void
add_remap(clone_state * state,void * nptr,const void * ptr)103 add_remap(clone_state *state, void *nptr, const void *ptr)
104 {
105 _mesa_hash_table_insert(state->remap_table, ptr, nptr);
106 }
107
108 static void *
remap_local(clone_state * state,const void * ptr)109 remap_local(clone_state *state, const void *ptr)
110 {
111 return _lookup_ptr(state, ptr, false);
112 }
113
114 static void *
remap_global(clone_state * state,const void * ptr)115 remap_global(clone_state *state, const void *ptr)
116 {
117 return _lookup_ptr(state, ptr, true);
118 }
119
120 static nir_register *
remap_reg(clone_state * state,const nir_register * reg)121 remap_reg(clone_state *state, const nir_register *reg)
122 {
123 return _lookup_ptr(state, reg, false);
124 }
125
126 static nir_variable *
remap_var(clone_state * state,const nir_variable * var)127 remap_var(clone_state *state, const nir_variable *var)
128 {
129 return _lookup_ptr(state, var, nir_variable_is_global(var));
130 }
131
132 nir_constant *
nir_constant_clone(const nir_constant * c,nir_variable * nvar)133 nir_constant_clone(const nir_constant *c, nir_variable *nvar)
134 {
135 nir_constant *nc = ralloc(nvar, nir_constant);
136
137 memcpy(nc->values, c->values, sizeof(nc->values));
138 nc->num_elements = c->num_elements;
139 nc->elements = ralloc_array(nvar, nir_constant *, c->num_elements);
140 for (unsigned i = 0; i < c->num_elements; i++) {
141 nc->elements[i] = nir_constant_clone(c->elements[i], nvar);
142 }
143
144 return nc;
145 }
146
147 /* NOTE: for cloning nir_variables, bypass nir_variable_create to avoid
148 * having to deal with locals and globals separately:
149 */
150 nir_variable *
nir_variable_clone(const nir_variable * var,nir_shader * shader)151 nir_variable_clone(const nir_variable *var, nir_shader *shader)
152 {
153 nir_variable *nvar = rzalloc(shader, nir_variable);
154
155 nvar->type = var->type;
156 nvar->name = ralloc_strdup(nvar, var->name);
157 nvar->data = var->data;
158 nvar->num_state_slots = var->num_state_slots;
159 if (var->num_state_slots) {
160 nvar->state_slots = ralloc_array(nvar, nir_state_slot, var->num_state_slots);
161 memcpy(nvar->state_slots, var->state_slots,
162 var->num_state_slots * sizeof(nir_state_slot));
163 }
164 if (var->constant_initializer) {
165 nvar->constant_initializer =
166 nir_constant_clone(var->constant_initializer, nvar);
167 }
168 nvar->interface_type = var->interface_type;
169
170 nvar->num_members = var->num_members;
171 if (var->num_members) {
172 nvar->members = ralloc_array(nvar, struct nir_variable_data,
173 var->num_members);
174 memcpy(nvar->members, var->members,
175 var->num_members * sizeof(*var->members));
176 }
177
178 return nvar;
179 }
180
181 static nir_variable *
clone_variable(clone_state * state,const nir_variable * var)182 clone_variable(clone_state *state, const nir_variable *var)
183 {
184 nir_variable *nvar = nir_variable_clone(var, state->ns);
185 add_remap(state, nvar, var);
186
187 return nvar;
188 }
189
190 /* clone list of nir_variable: */
191 static void
clone_var_list(clone_state * state,struct exec_list * dst,const struct exec_list * list)192 clone_var_list(clone_state *state, struct exec_list *dst,
193 const struct exec_list *list)
194 {
195 exec_list_make_empty(dst);
196 foreach_list_typed(nir_variable, var, node, list) {
197 nir_variable *nvar = clone_variable(state, var);
198 exec_list_push_tail(dst, &nvar->node);
199 }
200 }
201
202 /* NOTE: for cloning nir_registers, bypass nir_global/local_reg_create()
203 * to avoid having to deal with locals and globals separately:
204 */
205 static nir_register *
clone_register(clone_state * state,const nir_register * reg)206 clone_register(clone_state *state, const nir_register *reg)
207 {
208 nir_register *nreg = rzalloc(state->ns, nir_register);
209 add_remap(state, nreg, reg);
210
211 nreg->num_components = reg->num_components;
212 nreg->bit_size = reg->bit_size;
213 nreg->num_array_elems = reg->num_array_elems;
214 nreg->index = reg->index;
215 nreg->name = ralloc_strdup(nreg, reg->name);
216
217 /* reconstructing uses/defs/if_uses handled by nir_instr_insert() */
218 list_inithead(&nreg->uses);
219 list_inithead(&nreg->defs);
220 list_inithead(&nreg->if_uses);
221
222 return nreg;
223 }
224
225 /* clone list of nir_register: */
226 static void
clone_reg_list(clone_state * state,struct exec_list * dst,const struct exec_list * list)227 clone_reg_list(clone_state *state, struct exec_list *dst,
228 const struct exec_list *list)
229 {
230 exec_list_make_empty(dst);
231 foreach_list_typed(nir_register, reg, node, list) {
232 nir_register *nreg = clone_register(state, reg);
233 exec_list_push_tail(dst, &nreg->node);
234 }
235 }
236
237 static void
__clone_src(clone_state * state,void * ninstr_or_if,nir_src * nsrc,const nir_src * src)238 __clone_src(clone_state *state, void *ninstr_or_if,
239 nir_src *nsrc, const nir_src *src)
240 {
241 nsrc->is_ssa = src->is_ssa;
242 if (src->is_ssa) {
243 nsrc->ssa = remap_local(state, src->ssa);
244 } else {
245 nsrc->reg.reg = remap_reg(state, src->reg.reg);
246 if (src->reg.indirect) {
247 nsrc->reg.indirect = ralloc(ninstr_or_if, nir_src);
248 __clone_src(state, ninstr_or_if, nsrc->reg.indirect, src->reg.indirect);
249 }
250 nsrc->reg.base_offset = src->reg.base_offset;
251 }
252 }
253
254 static void
__clone_dst(clone_state * state,nir_instr * ninstr,nir_dest * ndst,const nir_dest * dst)255 __clone_dst(clone_state *state, nir_instr *ninstr,
256 nir_dest *ndst, const nir_dest *dst)
257 {
258 ndst->is_ssa = dst->is_ssa;
259 if (dst->is_ssa) {
260 nir_ssa_dest_init(ninstr, ndst, dst->ssa.num_components,
261 dst->ssa.bit_size, dst->ssa.name);
262 if (likely(state->remap_table))
263 add_remap(state, &ndst->ssa, &dst->ssa);
264 } else {
265 ndst->reg.reg = remap_reg(state, dst->reg.reg);
266 if (dst->reg.indirect) {
267 ndst->reg.indirect = ralloc(ninstr, nir_src);
268 __clone_src(state, ninstr, ndst->reg.indirect, dst->reg.indirect);
269 }
270 ndst->reg.base_offset = dst->reg.base_offset;
271 }
272 }
273
274 static nir_alu_instr *
clone_alu(clone_state * state,const nir_alu_instr * alu)275 clone_alu(clone_state *state, const nir_alu_instr *alu)
276 {
277 nir_alu_instr *nalu = nir_alu_instr_create(state->ns, alu->op);
278 nalu->exact = alu->exact;
279 nalu->no_signed_wrap = alu->no_signed_wrap;
280 nalu->no_unsigned_wrap = alu->no_unsigned_wrap;
281
282 __clone_dst(state, &nalu->instr, &nalu->dest.dest, &alu->dest.dest);
283 nalu->dest.saturate = alu->dest.saturate;
284 nalu->dest.write_mask = alu->dest.write_mask;
285
286 for (unsigned i = 0; i < nir_op_infos[alu->op].num_inputs; i++) {
287 __clone_src(state, &nalu->instr, &nalu->src[i].src, &alu->src[i].src);
288 nalu->src[i].negate = alu->src[i].negate;
289 nalu->src[i].abs = alu->src[i].abs;
290 memcpy(nalu->src[i].swizzle, alu->src[i].swizzle,
291 sizeof(nalu->src[i].swizzle));
292 }
293
294 return nalu;
295 }
296
297 nir_alu_instr *
nir_alu_instr_clone(nir_shader * shader,const nir_alu_instr * orig)298 nir_alu_instr_clone(nir_shader *shader, const nir_alu_instr *orig)
299 {
300 clone_state state = {
301 .allow_remap_fallback = true,
302 .ns = shader,
303 };
304 return clone_alu(&state, orig);
305 }
306
307 static nir_deref_instr *
clone_deref_instr(clone_state * state,const nir_deref_instr * deref)308 clone_deref_instr(clone_state *state, const nir_deref_instr *deref)
309 {
310 nir_deref_instr *nderef =
311 nir_deref_instr_create(state->ns, deref->deref_type);
312
313 __clone_dst(state, &nderef->instr, &nderef->dest, &deref->dest);
314
315 nderef->modes = deref->modes;
316 nderef->type = deref->type;
317
318 if (deref->deref_type == nir_deref_type_var) {
319 nderef->var = remap_var(state, deref->var);
320 return nderef;
321 }
322
323 __clone_src(state, &nderef->instr, &nderef->parent, &deref->parent);
324
325 switch (deref->deref_type) {
326 case nir_deref_type_struct:
327 nderef->strct.index = deref->strct.index;
328 break;
329
330 case nir_deref_type_array:
331 case nir_deref_type_ptr_as_array:
332 __clone_src(state, &nderef->instr,
333 &nderef->arr.index, &deref->arr.index);
334 break;
335
336 case nir_deref_type_array_wildcard:
337 /* Nothing to do */
338 break;
339
340 case nir_deref_type_cast:
341 nderef->cast.ptr_stride = deref->cast.ptr_stride;
342 nderef->cast.align_mul = deref->cast.align_mul;
343 nderef->cast.align_offset = deref->cast.align_offset;
344 break;
345
346 default:
347 unreachable("Invalid instruction deref type");
348 }
349
350 return nderef;
351 }
352
353 static nir_intrinsic_instr *
clone_intrinsic(clone_state * state,const nir_intrinsic_instr * itr)354 clone_intrinsic(clone_state *state, const nir_intrinsic_instr *itr)
355 {
356 nir_intrinsic_instr *nitr =
357 nir_intrinsic_instr_create(state->ns, itr->intrinsic);
358
359 unsigned num_srcs = nir_intrinsic_infos[itr->intrinsic].num_srcs;
360
361 if (nir_intrinsic_infos[itr->intrinsic].has_dest)
362 __clone_dst(state, &nitr->instr, &nitr->dest, &itr->dest);
363
364 nitr->num_components = itr->num_components;
365 memcpy(nitr->const_index, itr->const_index, sizeof(nitr->const_index));
366
367 for (unsigned i = 0; i < num_srcs; i++)
368 __clone_src(state, &nitr->instr, &nitr->src[i], &itr->src[i]);
369
370 return nitr;
371 }
372
373 static nir_load_const_instr *
clone_load_const(clone_state * state,const nir_load_const_instr * lc)374 clone_load_const(clone_state *state, const nir_load_const_instr *lc)
375 {
376 nir_load_const_instr *nlc =
377 nir_load_const_instr_create(state->ns, lc->def.num_components,
378 lc->def.bit_size);
379
380 memcpy(&nlc->value, &lc->value, sizeof(*nlc->value) * lc->def.num_components);
381
382 add_remap(state, &nlc->def, &lc->def);
383
384 return nlc;
385 }
386
387 static nir_ssa_undef_instr *
clone_ssa_undef(clone_state * state,const nir_ssa_undef_instr * sa)388 clone_ssa_undef(clone_state *state, const nir_ssa_undef_instr *sa)
389 {
390 nir_ssa_undef_instr *nsa =
391 nir_ssa_undef_instr_create(state->ns, sa->def.num_components,
392 sa->def.bit_size);
393
394 add_remap(state, &nsa->def, &sa->def);
395
396 return nsa;
397 }
398
399 static nir_tex_instr *
clone_tex(clone_state * state,const nir_tex_instr * tex)400 clone_tex(clone_state *state, const nir_tex_instr *tex)
401 {
402 nir_tex_instr *ntex = nir_tex_instr_create(state->ns, tex->num_srcs);
403
404 ntex->sampler_dim = tex->sampler_dim;
405 ntex->dest_type = tex->dest_type;
406 ntex->op = tex->op;
407 __clone_dst(state, &ntex->instr, &ntex->dest, &tex->dest);
408 for (unsigned i = 0; i < ntex->num_srcs; i++) {
409 ntex->src[i].src_type = tex->src[i].src_type;
410 __clone_src(state, &ntex->instr, &ntex->src[i].src, &tex->src[i].src);
411 }
412 ntex->coord_components = tex->coord_components;
413 ntex->is_array = tex->is_array;
414 ntex->is_shadow = tex->is_shadow;
415 ntex->is_new_style_shadow = tex->is_new_style_shadow;
416 ntex->component = tex->component;
417 memcpy(ntex->tg4_offsets, tex->tg4_offsets, sizeof(tex->tg4_offsets));
418
419 ntex->texture_index = tex->texture_index;
420 ntex->sampler_index = tex->sampler_index;
421
422 ntex->texture_non_uniform = tex->texture_non_uniform;
423 ntex->sampler_non_uniform = tex->sampler_non_uniform;
424
425 return ntex;
426 }
427
428 static nir_phi_instr *
clone_phi(clone_state * state,const nir_phi_instr * phi,nir_block * nblk)429 clone_phi(clone_state *state, const nir_phi_instr *phi, nir_block *nblk)
430 {
431 nir_phi_instr *nphi = nir_phi_instr_create(state->ns);
432
433 __clone_dst(state, &nphi->instr, &nphi->dest, &phi->dest);
434
435 /* Cloning a phi node is a bit different from other instructions. The
436 * sources of phi instructions are the only time where we can use an SSA
437 * def before it is defined. In order to handle this, we just copy over
438 * the sources from the old phi instruction directly and then fix them up
439 * in a second pass once all the instrutions in the function have been
440 * properly cloned.
441 *
442 * In order to ensure that the copied sources (which are the same as the
443 * old phi instruction's sources for now) don't get inserted into the old
444 * shader's use-def lists, we have to add the phi instruction *before* we
445 * set up its sources.
446 */
447 nir_instr_insert_after_block(nblk, &nphi->instr);
448
449 foreach_list_typed(nir_phi_src, src, node, &phi->srcs) {
450 nir_phi_src *nsrc = ralloc(nphi, nir_phi_src);
451
452 /* Just copy the old source for now. */
453 memcpy(nsrc, src, sizeof(*src));
454
455 /* Since we're not letting nir_insert_instr handle use/def stuff for us,
456 * we have to set the parent_instr manually. It doesn't really matter
457 * when we do it, so we might as well do it here.
458 */
459 nsrc->src.parent_instr = &nphi->instr;
460
461 /* Stash it in the list of phi sources. We'll walk this list and fix up
462 * sources at the very end of clone_function_impl.
463 */
464 list_add(&nsrc->src.use_link, &state->phi_srcs);
465
466 exec_list_push_tail(&nphi->srcs, &nsrc->node);
467 }
468
469 return nphi;
470 }
471
472 static nir_jump_instr *
clone_jump(clone_state * state,const nir_jump_instr * jmp)473 clone_jump(clone_state *state, const nir_jump_instr *jmp)
474 {
475 /* These aren't handled because they require special block linking */
476 assert(jmp->type != nir_jump_goto && jmp->type != nir_jump_goto_if);
477
478 nir_jump_instr *njmp = nir_jump_instr_create(state->ns, jmp->type);
479
480 return njmp;
481 }
482
483 static nir_call_instr *
clone_call(clone_state * state,const nir_call_instr * call)484 clone_call(clone_state *state, const nir_call_instr *call)
485 {
486 nir_function *ncallee = remap_global(state, call->callee);
487 nir_call_instr *ncall = nir_call_instr_create(state->ns, ncallee);
488
489 for (unsigned i = 0; i < ncall->num_params; i++)
490 __clone_src(state, ncall, &ncall->params[i], &call->params[i]);
491
492 return ncall;
493 }
494
495 static nir_instr *
clone_instr(clone_state * state,const nir_instr * instr)496 clone_instr(clone_state *state, const nir_instr *instr)
497 {
498 switch (instr->type) {
499 case nir_instr_type_alu:
500 return &clone_alu(state, nir_instr_as_alu(instr))->instr;
501 case nir_instr_type_deref:
502 return &clone_deref_instr(state, nir_instr_as_deref(instr))->instr;
503 case nir_instr_type_intrinsic:
504 return &clone_intrinsic(state, nir_instr_as_intrinsic(instr))->instr;
505 case nir_instr_type_load_const:
506 return &clone_load_const(state, nir_instr_as_load_const(instr))->instr;
507 case nir_instr_type_ssa_undef:
508 return &clone_ssa_undef(state, nir_instr_as_ssa_undef(instr))->instr;
509 case nir_instr_type_tex:
510 return &clone_tex(state, nir_instr_as_tex(instr))->instr;
511 case nir_instr_type_phi:
512 unreachable("Cannot clone phis with clone_instr");
513 case nir_instr_type_jump:
514 return &clone_jump(state, nir_instr_as_jump(instr))->instr;
515 case nir_instr_type_call:
516 return &clone_call(state, nir_instr_as_call(instr))->instr;
517 case nir_instr_type_parallel_copy:
518 unreachable("Cannot clone parallel copies");
519 default:
520 unreachable("bad instr type");
521 return NULL;
522 }
523 }
524
525 nir_instr *
nir_instr_clone(nir_shader * shader,const nir_instr * orig)526 nir_instr_clone(nir_shader *shader, const nir_instr *orig)
527 {
528 clone_state state = {
529 .allow_remap_fallback = true,
530 .ns = shader,
531 };
532 return clone_instr(&state, orig);
533 }
534
535 static nir_block *
clone_block(clone_state * state,struct exec_list * cf_list,const nir_block * blk)536 clone_block(clone_state *state, struct exec_list *cf_list, const nir_block *blk)
537 {
538 /* Don't actually create a new block. Just use the one from the tail of
539 * the list. NIR guarantees that the tail of the list is a block and that
540 * no two blocks are side-by-side in the IR; It should be empty.
541 */
542 nir_block *nblk =
543 exec_node_data(nir_block, exec_list_get_tail(cf_list), cf_node.node);
544 assert(nblk->cf_node.type == nir_cf_node_block);
545 assert(exec_list_is_empty(&nblk->instr_list));
546
547 /* We need this for phi sources */
548 add_remap(state, nblk, blk);
549
550 nir_foreach_instr(instr, blk) {
551 if (instr->type == nir_instr_type_phi) {
552 /* Phi instructions are a bit of a special case when cloning because
553 * we don't want inserting the instruction to automatically handle
554 * use/defs for us. Instead, we need to wait until all the
555 * blocks/instructions are in so that we can set their sources up.
556 */
557 clone_phi(state, nir_instr_as_phi(instr), nblk);
558 } else {
559 nir_instr *ninstr = clone_instr(state, instr);
560 nir_instr_insert_after_block(nblk, ninstr);
561 }
562 }
563
564 return nblk;
565 }
566
567 static void
568 clone_cf_list(clone_state *state, struct exec_list *dst,
569 const struct exec_list *list);
570
571 static nir_if *
clone_if(clone_state * state,struct exec_list * cf_list,const nir_if * i)572 clone_if(clone_state *state, struct exec_list *cf_list, const nir_if *i)
573 {
574 nir_if *ni = nir_if_create(state->ns);
575 ni->control = i->control;
576
577 __clone_src(state, ni, &ni->condition, &i->condition);
578
579 nir_cf_node_insert_end(cf_list, &ni->cf_node);
580
581 clone_cf_list(state, &ni->then_list, &i->then_list);
582 clone_cf_list(state, &ni->else_list, &i->else_list);
583
584 return ni;
585 }
586
587 static nir_loop *
clone_loop(clone_state * state,struct exec_list * cf_list,const nir_loop * loop)588 clone_loop(clone_state *state, struct exec_list *cf_list, const nir_loop *loop)
589 {
590 nir_loop *nloop = nir_loop_create(state->ns);
591 nloop->control = loop->control;
592 nloop->partially_unrolled = loop->partially_unrolled;
593
594 nir_cf_node_insert_end(cf_list, &nloop->cf_node);
595
596 clone_cf_list(state, &nloop->body, &loop->body);
597
598 return nloop;
599 }
600
601 /* clone list of nir_cf_node: */
602 static void
clone_cf_list(clone_state * state,struct exec_list * dst,const struct exec_list * list)603 clone_cf_list(clone_state *state, struct exec_list *dst,
604 const struct exec_list *list)
605 {
606 foreach_list_typed(nir_cf_node, cf, node, list) {
607 switch (cf->type) {
608 case nir_cf_node_block:
609 clone_block(state, dst, nir_cf_node_as_block(cf));
610 break;
611 case nir_cf_node_if:
612 clone_if(state, dst, nir_cf_node_as_if(cf));
613 break;
614 case nir_cf_node_loop:
615 clone_loop(state, dst, nir_cf_node_as_loop(cf));
616 break;
617 default:
618 unreachable("bad cf type");
619 }
620 }
621 }
622
623 /* After we've cloned almost everything, we have to walk the list of phi
624 * sources and fix them up. Thanks to loops, the block and SSA value for a
625 * phi source may not be defined when we first encounter it. Instead, we
626 * add it to the phi_srcs list and we fix it up here.
627 */
628 static void
fixup_phi_srcs(clone_state * state)629 fixup_phi_srcs(clone_state *state)
630 {
631 list_for_each_entry_safe(nir_phi_src, src, &state->phi_srcs, src.use_link) {
632 src->pred = remap_local(state, src->pred);
633
634 /* Remove from this list */
635 list_del(&src->src.use_link);
636
637 if (src->src.is_ssa) {
638 src->src.ssa = remap_local(state, src->src.ssa);
639 list_addtail(&src->src.use_link, &src->src.ssa->uses);
640 } else {
641 src->src.reg.reg = remap_reg(state, src->src.reg.reg);
642 list_addtail(&src->src.use_link, &src->src.reg.reg->uses);
643 }
644 }
645 assert(list_is_empty(&state->phi_srcs));
646 }
647
648 void
nir_cf_list_clone(nir_cf_list * dst,nir_cf_list * src,nir_cf_node * parent,struct hash_table * remap_table)649 nir_cf_list_clone(nir_cf_list *dst, nir_cf_list *src, nir_cf_node *parent,
650 struct hash_table *remap_table)
651 {
652 exec_list_make_empty(&dst->list);
653 dst->impl = src->impl;
654
655 if (exec_list_is_empty(&src->list))
656 return;
657
658 clone_state state;
659 init_clone_state(&state, remap_table, false, true);
660
661 /* We use the same shader */
662 state.ns = src->impl->function->shader;
663
664 /* The control-flow code assumes that the list of cf_nodes always starts
665 * and ends with a block. We start by adding an empty block.
666 */
667 nir_block *nblk = nir_block_create(state.ns);
668 nblk->cf_node.parent = parent;
669 exec_list_push_tail(&dst->list, &nblk->cf_node.node);
670
671 clone_cf_list(&state, &dst->list, &src->list);
672
673 fixup_phi_srcs(&state);
674
675 if (!remap_table)
676 free_clone_state(&state);
677 }
678
679 static nir_function_impl *
clone_function_impl(clone_state * state,const nir_function_impl * fi)680 clone_function_impl(clone_state *state, const nir_function_impl *fi)
681 {
682 nir_function_impl *nfi = nir_function_impl_create_bare(state->ns);
683
684 clone_var_list(state, &nfi->locals, &fi->locals);
685 clone_reg_list(state, &nfi->registers, &fi->registers);
686 nfi->reg_alloc = fi->reg_alloc;
687
688 assert(list_is_empty(&state->phi_srcs));
689
690 clone_cf_list(state, &nfi->body, &fi->body);
691
692 fixup_phi_srcs(state);
693
694 /* All metadata is invalidated in the cloning process */
695 nfi->valid_metadata = 0;
696
697 return nfi;
698 }
699
700 nir_function_impl *
nir_function_impl_clone(nir_shader * shader,const nir_function_impl * fi)701 nir_function_impl_clone(nir_shader *shader, const nir_function_impl *fi)
702 {
703 clone_state state;
704 init_clone_state(&state, NULL, false, false);
705
706 state.ns = shader;
707
708 nir_function_impl *nfi = clone_function_impl(&state, fi);
709
710 free_clone_state(&state);
711
712 return nfi;
713 }
714
715 static nir_function *
clone_function(clone_state * state,const nir_function * fxn,nir_shader * ns)716 clone_function(clone_state *state, const nir_function *fxn, nir_shader *ns)
717 {
718 assert(ns == state->ns);
719 nir_function *nfxn = nir_function_create(ns, fxn->name);
720
721 /* Needed for call instructions */
722 add_remap(state, nfxn, fxn);
723
724 nfxn->num_params = fxn->num_params;
725 if (fxn->num_params) {
726 nfxn->params = ralloc_array(state->ns, nir_parameter, fxn->num_params);
727 memcpy(nfxn->params, fxn->params, sizeof(nir_parameter) * fxn->num_params);
728 }
729 nfxn->is_entrypoint = fxn->is_entrypoint;
730
731 /* At first glance, it looks like we should clone the function_impl here.
732 * However, call instructions need to be able to reference at least the
733 * function and those will get processed as we clone the function_impls.
734 * We stop here and do function_impls as a second pass.
735 */
736
737 return nfxn;
738 }
739
740 nir_shader *
nir_shader_clone(void * mem_ctx,const nir_shader * s)741 nir_shader_clone(void *mem_ctx, const nir_shader *s)
742 {
743 clone_state state;
744 init_clone_state(&state, NULL, true, false);
745
746 nir_shader *ns = nir_shader_create(mem_ctx, s->info.stage, s->options, NULL);
747 state.ns = ns;
748
749 clone_var_list(&state, &ns->variables, &s->variables);
750
751 /* Go through and clone functions */
752 foreach_list_typed(nir_function, fxn, node, &s->functions)
753 clone_function(&state, fxn, ns);
754
755 /* Only after all functions are cloned can we clone the actual function
756 * implementations. This is because nir_call_instrs need to reference the
757 * functions of other functions and we don't know what order the functions
758 * will have in the list.
759 */
760 nir_foreach_function(fxn, s) {
761 nir_function *nfxn = remap_global(&state, fxn);
762 nfxn->impl = clone_function_impl(&state, fxn->impl);
763 nfxn->impl->function = nfxn;
764 }
765
766 ns->info = s->info;
767 ns->info.name = ralloc_strdup(ns, ns->info.name);
768 if (ns->info.label)
769 ns->info.label = ralloc_strdup(ns, ns->info.label);
770
771 ns->num_inputs = s->num_inputs;
772 ns->num_uniforms = s->num_uniforms;
773 ns->num_outputs = s->num_outputs;
774 ns->shared_size = s->shared_size;
775 ns->scratch_size = s->scratch_size;
776
777 ns->constant_data_size = s->constant_data_size;
778 if (s->constant_data_size > 0) {
779 ns->constant_data = ralloc_size(ns, s->constant_data_size);
780 memcpy(ns->constant_data, s->constant_data, s->constant_data_size);
781 }
782
783 free_clone_state(&state);
784
785 return ns;
786 }
787
788 /** Overwrites dst and replaces its contents with src
789 *
790 * Everything ralloc parented to dst and src itself (but not its children)
791 * will be freed.
792 *
793 * This should only be used by test code which needs to swap out shaders with
794 * a cloned or deserialized version.
795 */
796 void
nir_shader_replace(nir_shader * dst,nir_shader * src)797 nir_shader_replace(nir_shader *dst, nir_shader *src)
798 {
799 /* Delete all of dest's ralloc children */
800 void *dead_ctx = ralloc_context(NULL);
801 ralloc_adopt(dead_ctx, dst);
802 ralloc_free(dead_ctx);
803
804 /* Re-parent all of src's ralloc children to dst */
805 ralloc_adopt(dst, src);
806
807 memcpy(dst, src, sizeof(*dst));
808
809 /* We have to move all the linked lists over separately because we need the
810 * pointers in the list elements to point to the lists in dst and not src.
811 */
812 exec_list_move_nodes_to(&src->variables, &dst->variables);
813
814 /* Now move the functions over. This takes a tiny bit more work */
815 exec_list_move_nodes_to(&src->functions, &dst->functions);
816 nir_foreach_function(function, dst)
817 function->shader = dst;
818
819 ralloc_free(src);
820 }
821