1 /*
2 * Copyright © 2010 Intel Corporation
3 *
4 * Permission is hereby granted, free of charge, to any person obtaining a
5 * copy of this software and associated documentation files (the "Software"),
6 * to deal in the Software without restriction, including without limitation
7 * the rights to use, copy, modify, merge, publish, distribute, sublicense,
8 * and/or sell copies of the Software, and to permit persons to whom the
9 * Software is furnished to do so, subject to the following conditions:
10 *
11 * The above copyright notice and this permission notice (including the next
12 * paragraph) shall be included in all copies or substantial portions of the
13 * Software.
14 *
15 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
16 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
17 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
18 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
19 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
20 * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
21 * DEALINGS IN THE SOFTWARE.
22 */
23
24 /**
25 * \file opt_algebraic.cpp
26 *
27 * Takes advantage of association, commutivity, and other algebraic
28 * properties to simplify expressions.
29 */
30
31 #include "ir.h"
32 #include "ir_visitor.h"
33 #include "ir_rvalue_visitor.h"
34 #include "ir_optimization.h"
35 #include "ir_builder.h"
36 #include "compiler/glsl_types.h"
37 #include "main/consts_exts.h"
38
39 using namespace ir_builder;
40
41 namespace {
42
43 /**
44 * Visitor class for replacing expressions with ir_constant values.
45 */
46
47 class ir_algebraic_visitor : public ir_rvalue_visitor {
48 public:
ir_algebraic_visitor(bool native_integers,const struct gl_shader_compiler_options * options)49 ir_algebraic_visitor(bool native_integers,
50 const struct gl_shader_compiler_options *options)
51 : options(options)
52 {
53 this->progress = false;
54 this->mem_ctx = NULL;
55 this->native_integers = native_integers;
56 }
57
~ir_algebraic_visitor()58 virtual ~ir_algebraic_visitor()
59 {
60 }
61
62 virtual ir_visitor_status visit_enter(ir_assignment *ir);
63
64 ir_rvalue *handle_expression(ir_expression *ir);
65 void handle_rvalue(ir_rvalue **rvalue);
66 bool reassociate_constant(ir_expression *ir1,
67 int const_index,
68 ir_constant *constant,
69 ir_expression *ir2);
70 void reassociate_operands(ir_expression *ir1,
71 int op1,
72 ir_expression *ir2,
73 int op2);
74 ir_rvalue *swizzle_if_required(ir_expression *expr,
75 ir_rvalue *operand);
76
77 const struct gl_shader_compiler_options *options;
78 void *mem_ctx;
79
80 bool native_integers;
81 bool progress;
82 };
83
84 } /* unnamed namespace */
85
86 ir_visitor_status
visit_enter(ir_assignment * ir)87 ir_algebraic_visitor::visit_enter(ir_assignment *ir)
88 {
89 ir_variable *var = ir->lhs->variable_referenced();
90 if (var->data.invariant || var->data.precise) {
91 /* If we're assigning to an invariant or precise variable, just bail.
92 * Most of the algebraic optimizations aren't precision-safe.
93 *
94 * FINISHME: Find out which optimizations are precision-safe and enable
95 * then only for invariant or precise trees.
96 */
97 return visit_continue_with_parent;
98 } else {
99 return visit_continue;
100 }
101 }
102
103 static inline bool
is_vec_zero(ir_constant * ir)104 is_vec_zero(ir_constant *ir)
105 {
106 return (ir == NULL) ? false : ir->is_zero();
107 }
108
109 static inline bool
is_vec_one(ir_constant * ir)110 is_vec_one(ir_constant *ir)
111 {
112 return (ir == NULL) ? false : ir->is_one();
113 }
114
115 static inline bool
is_vec_two(ir_constant * ir)116 is_vec_two(ir_constant *ir)
117 {
118 return (ir == NULL) ? false : ir->is_value(2.0, 2);
119 }
120
121 static inline bool
is_vec_four(ir_constant * ir)122 is_vec_four(ir_constant *ir)
123 {
124 return (ir == NULL) ? false : ir->is_value(4.0, 4);
125 }
126
127 static inline bool
is_vec_negative_one(ir_constant * ir)128 is_vec_negative_one(ir_constant *ir)
129 {
130 return (ir == NULL) ? false : ir->is_negative_one();
131 }
132
133 static inline bool
is_valid_vec_const(ir_constant * ir)134 is_valid_vec_const(ir_constant *ir)
135 {
136 if (ir == NULL)
137 return false;
138
139 if (!ir->type->is_scalar() && !ir->type->is_vector())
140 return false;
141
142 return true;
143 }
144
145 static inline bool
is_less_than_one(ir_constant * ir)146 is_less_than_one(ir_constant *ir)
147 {
148 assert(ir->type->is_float());
149
150 if (!is_valid_vec_const(ir))
151 return false;
152
153 unsigned component = 0;
154 for (int c = 0; c < ir->type->vector_elements; c++) {
155 if (ir->get_float_component(c) < 1.0f)
156 component++;
157 }
158
159 return (component == ir->type->vector_elements);
160 }
161
162 static inline bool
is_greater_than_zero(ir_constant * ir)163 is_greater_than_zero(ir_constant *ir)
164 {
165 assert(ir->type->is_float());
166
167 if (!is_valid_vec_const(ir))
168 return false;
169
170 unsigned component = 0;
171 for (int c = 0; c < ir->type->vector_elements; c++) {
172 if (ir->get_float_component(c) > 0.0f)
173 component++;
174 }
175
176 return (component == ir->type->vector_elements);
177 }
178
179 static void
update_type(ir_expression * ir)180 update_type(ir_expression *ir)
181 {
182 if (ir->operands[0]->type->is_vector())
183 ir->type = ir->operands[0]->type;
184 else
185 ir->type = ir->operands[1]->type;
186 }
187
188 /* Recognize (v.x + v.y) + (v.z + v.w) as dot(v, 1.0) */
189 static ir_expression *
try_replace_with_dot(ir_expression * expr0,ir_expression * expr1,void * mem_ctx)190 try_replace_with_dot(ir_expression *expr0, ir_expression *expr1, void *mem_ctx)
191 {
192 if (expr0 && expr0->operation == ir_binop_add &&
193 expr0->type->is_float() &&
194 expr1 && expr1->operation == ir_binop_add &&
195 expr1->type->is_float()) {
196 ir_swizzle *x = expr0->operands[0]->as_swizzle();
197 ir_swizzle *y = expr0->operands[1]->as_swizzle();
198 ir_swizzle *z = expr1->operands[0]->as_swizzle();
199 ir_swizzle *w = expr1->operands[1]->as_swizzle();
200
201 if (!x || x->mask.num_components != 1 ||
202 !y || y->mask.num_components != 1 ||
203 !z || z->mask.num_components != 1 ||
204 !w || w->mask.num_components != 1) {
205 return NULL;
206 }
207
208 bool swiz_seen[4] = {false, false, false, false};
209 swiz_seen[x->mask.x] = true;
210 swiz_seen[y->mask.x] = true;
211 swiz_seen[z->mask.x] = true;
212 swiz_seen[w->mask.x] = true;
213
214 if (!swiz_seen[0] || !swiz_seen[1] ||
215 !swiz_seen[2] || !swiz_seen[3]) {
216 return NULL;
217 }
218
219 if (x->val->equals(y->val) &&
220 x->val->equals(z->val) &&
221 x->val->equals(w->val)) {
222 return dot(x->val, new(mem_ctx) ir_constant(1.0f, 4));
223 }
224 }
225 return NULL;
226 }
227
228 void
reassociate_operands(ir_expression * ir1,int op1,ir_expression * ir2,int op2)229 ir_algebraic_visitor::reassociate_operands(ir_expression *ir1,
230 int op1,
231 ir_expression *ir2,
232 int op2)
233 {
234 ir_rvalue *temp = ir2->operands[op2];
235 ir2->operands[op2] = ir1->operands[op1];
236 ir1->operands[op1] = temp;
237
238 /* Update the type of ir2. The type of ir1 won't have changed --
239 * base types matched, and at least one of the operands of the 2
240 * binops is still a vector if any of them were.
241 */
242 update_type(ir2);
243
244 this->progress = true;
245 }
246
247 /**
248 * Reassociates a constant down a tree of adds or multiplies.
249 *
250 * Consider (2 * (a * (b * 0.5))). We want to end up with a * b.
251 */
252 bool
reassociate_constant(ir_expression * ir1,int const_index,ir_constant * constant,ir_expression * ir2)253 ir_algebraic_visitor::reassociate_constant(ir_expression *ir1, int const_index,
254 ir_constant *constant,
255 ir_expression *ir2)
256 {
257 if (!ir2 || ir1->operation != ir2->operation)
258 return false;
259
260 /* Don't want to even think about matrices. */
261 if (ir1->operands[0]->type->is_matrix() ||
262 ir1->operands[1]->type->is_matrix() ||
263 ir2->operands[0]->type->is_matrix() ||
264 ir2->operands[1]->type->is_matrix())
265 return false;
266
267 void *mem_ctx = ralloc_parent(ir2);
268
269 ir_constant *ir2_const[2];
270 ir2_const[0] = ir2->operands[0]->constant_expression_value(mem_ctx);
271 ir2_const[1] = ir2->operands[1]->constant_expression_value(mem_ctx);
272
273 if (ir2_const[0] && ir2_const[1])
274 return false;
275
276 if (ir2_const[0]) {
277 reassociate_operands(ir1, const_index, ir2, 1);
278 return true;
279 } else if (ir2_const[1]) {
280 reassociate_operands(ir1, const_index, ir2, 0);
281 return true;
282 }
283
284 if (reassociate_constant(ir1, const_index, constant,
285 ir2->operands[0]->as_expression())) {
286 update_type(ir2);
287 return true;
288 }
289
290 if (reassociate_constant(ir1, const_index, constant,
291 ir2->operands[1]->as_expression())) {
292 update_type(ir2);
293 return true;
294 }
295
296 return false;
297 }
298
299 /* When eliminating an expression and just returning one of its operands,
300 * we may need to swizzle that operand out to a vector if the expression was
301 * vector type.
302 */
303 ir_rvalue *
swizzle_if_required(ir_expression * expr,ir_rvalue * operand)304 ir_algebraic_visitor::swizzle_if_required(ir_expression *expr,
305 ir_rvalue *operand)
306 {
307 if (expr->type->is_vector() && operand->type->is_scalar()) {
308 return new(mem_ctx) ir_swizzle(operand, 0, 0, 0, 0,
309 expr->type->vector_elements);
310 } else
311 return operand;
312 }
313
314 ir_rvalue *
handle_expression(ir_expression * ir)315 ir_algebraic_visitor::handle_expression(ir_expression *ir)
316 {
317 ir_constant *op_const[4] = {NULL, NULL, NULL, NULL};
318 ir_expression *op_expr[4] = {NULL, NULL, NULL, NULL};
319
320 if (ir->operation == ir_binop_mul &&
321 ir->operands[0]->type->is_matrix() &&
322 ir->operands[1]->type->is_vector()) {
323 ir_expression *matrix_mul = ir->operands[0]->as_expression();
324
325 if (matrix_mul && matrix_mul->operation == ir_binop_mul &&
326 matrix_mul->operands[0]->type->is_matrix() &&
327 matrix_mul->operands[1]->type->is_matrix()) {
328
329 return mul(matrix_mul->operands[0],
330 mul(matrix_mul->operands[1], ir->operands[1]));
331 }
332 }
333
334 assert(ir->num_operands <= 4);
335 for (unsigned i = 0; i < ir->num_operands; i++) {
336 if (ir->operands[i]->type->is_matrix())
337 return ir;
338
339 op_const[i] =
340 ir->operands[i]->constant_expression_value(ralloc_parent(ir));
341 op_expr[i] = ir->operands[i]->as_expression();
342 }
343
344 if (this->mem_ctx == NULL)
345 this->mem_ctx = ralloc_parent(ir);
346
347 switch (ir->operation) {
348 case ir_unop_bit_not:
349 if (op_expr[0] && op_expr[0]->operation == ir_unop_bit_not)
350 return op_expr[0]->operands[0];
351 break;
352
353 case ir_unop_abs:
354 if (op_expr[0] == NULL)
355 break;
356
357 switch (op_expr[0]->operation) {
358 case ir_unop_abs:
359 case ir_unop_neg:
360 return abs(op_expr[0]->operands[0]);
361 default:
362 break;
363 }
364 break;
365
366 case ir_unop_neg:
367 if (op_expr[0] == NULL)
368 break;
369
370 if (op_expr[0]->operation == ir_unop_neg) {
371 return op_expr[0]->operands[0];
372 }
373 break;
374
375 case ir_unop_exp:
376 if (op_expr[0] == NULL)
377 break;
378
379 if (op_expr[0]->operation == ir_unop_log) {
380 return op_expr[0]->operands[0];
381 }
382 break;
383
384 case ir_unop_log:
385 if (op_expr[0] == NULL)
386 break;
387
388 if (op_expr[0]->operation == ir_unop_exp) {
389 return op_expr[0]->operands[0];
390 }
391 break;
392
393 case ir_unop_exp2:
394 if (op_expr[0] == NULL)
395 break;
396
397 if (op_expr[0]->operation == ir_unop_log2) {
398 return op_expr[0]->operands[0];
399 }
400
401 if (op_expr[0]->operation == ir_binop_mul) {
402 for (int log2_pos = 0; log2_pos < 2; log2_pos++) {
403 ir_expression *log2_expr =
404 op_expr[0]->operands[log2_pos]->as_expression();
405
406 if (log2_expr && log2_expr->operation == ir_unop_log2) {
407 return new(mem_ctx) ir_expression(ir_binop_pow,
408 ir->type,
409 log2_expr->operands[0],
410 op_expr[0]->operands[1 - log2_pos]);
411 }
412 }
413 }
414 break;
415
416 case ir_unop_log2:
417 if (op_expr[0] == NULL)
418 break;
419
420 if (op_expr[0]->operation == ir_unop_exp2) {
421 return op_expr[0]->operands[0];
422 }
423 break;
424
425 case ir_unop_f2i:
426 case ir_unop_f2u:
427 if (op_expr[0] && op_expr[0]->operation == ir_unop_trunc) {
428 return new(mem_ctx) ir_expression(ir->operation,
429 ir->type,
430 op_expr[0]->operands[0]);
431 }
432 break;
433
434 case ir_unop_logic_not: {
435 enum ir_expression_operation new_op = ir_unop_logic_not;
436
437 if (op_expr[0] == NULL)
438 break;
439
440 switch (op_expr[0]->operation) {
441 case ir_binop_less: new_op = ir_binop_gequal; break;
442 case ir_binop_gequal: new_op = ir_binop_less; break;
443 case ir_binop_equal: new_op = ir_binop_nequal; break;
444 case ir_binop_nequal: new_op = ir_binop_equal; break;
445 case ir_binop_all_equal: new_op = ir_binop_any_nequal; break;
446 case ir_binop_any_nequal: new_op = ir_binop_all_equal; break;
447
448 default:
449 /* The default case handler is here to silence a warning from GCC.
450 */
451 break;
452 }
453
454 if (new_op != ir_unop_logic_not) {
455 return new(mem_ctx) ir_expression(new_op,
456 ir->type,
457 op_expr[0]->operands[0],
458 op_expr[0]->operands[1]);
459 }
460
461 break;
462 }
463
464 case ir_unop_saturate:
465 if (op_expr[0] && op_expr[0]->operation == ir_binop_add) {
466 ir_expression *b2f_0 = op_expr[0]->operands[0]->as_expression();
467 ir_expression *b2f_1 = op_expr[0]->operands[1]->as_expression();
468
469 if (b2f_0 && b2f_0->operation == ir_unop_b2f &&
470 b2f_1 && b2f_1->operation == ir_unop_b2f) {
471 return b2f(logic_or(b2f_0->operands[0], b2f_1->operands[0]));
472 }
473 }
474 break;
475
476 /* This macro CANNOT use the do { } while(true) mechanism because
477 * then the breaks apply to the loop instead of the switch!
478 */
479 #define HANDLE_PACK_UNPACK_INVERSE(inverse_operation) \
480 { \
481 ir_expression *const op = ir->operands[0]->as_expression(); \
482 if (op == NULL) \
483 break; \
484 if (op->operation == (inverse_operation)) \
485 return op->operands[0]; \
486 break; \
487 }
488
489 case ir_unop_unpack_uint_2x32:
490 HANDLE_PACK_UNPACK_INVERSE(ir_unop_pack_uint_2x32);
491 case ir_unop_pack_uint_2x32:
492 HANDLE_PACK_UNPACK_INVERSE(ir_unop_unpack_uint_2x32);
493 case ir_unop_unpack_int_2x32:
494 HANDLE_PACK_UNPACK_INVERSE(ir_unop_pack_int_2x32);
495 case ir_unop_pack_int_2x32:
496 HANDLE_PACK_UNPACK_INVERSE(ir_unop_unpack_int_2x32);
497 case ir_unop_unpack_double_2x32:
498 HANDLE_PACK_UNPACK_INVERSE(ir_unop_pack_double_2x32);
499 case ir_unop_pack_double_2x32:
500 HANDLE_PACK_UNPACK_INVERSE(ir_unop_unpack_double_2x32);
501
502 #undef HANDLE_PACK_UNPACK_INVERSE
503
504 case ir_binop_add:
505 if (is_vec_zero(op_const[0]))
506 return ir->operands[1];
507 if (is_vec_zero(op_const[1]))
508 return ir->operands[0];
509
510 /* Replace (x + (-x)) with constant 0 */
511 for (int i = 0; i < 2; i++) {
512 if (op_expr[i]) {
513 if (op_expr[i]->operation == ir_unop_neg) {
514 ir_rvalue *other = ir->operands[(i + 1) % 2];
515 if (other && op_expr[i]->operands[0]->equals(other)) {
516 return ir_constant::zero(ir, ir->type);
517 }
518 }
519 }
520 }
521
522 /* Reassociate addition of constants so that we can do constant
523 * folding.
524 */
525 if (op_const[0] && !op_const[1])
526 reassociate_constant(ir, 0, op_const[0], op_expr[1]);
527 if (op_const[1] && !op_const[0])
528 reassociate_constant(ir, 1, op_const[1], op_expr[0]);
529
530 /* Recognize (v.x + v.y) + (v.z + v.w) as dot(v, 1.0) */
531 if (options->OptimizeForAOS) {
532 ir_expression *expr = try_replace_with_dot(op_expr[0], op_expr[1],
533 mem_ctx);
534 if (expr)
535 return expr;
536 }
537
538 /* Replace (-x + y) * a + x and commutative variations with lrp(x, y, a).
539 *
540 * (-x + y) * a + x
541 * (x * -a) + (y * a) + x
542 * x + (x * -a) + (y * a)
543 * x * (1 - a) + y * a
544 * lrp(x, y, a)
545 */
546 for (int mul_pos = 0; mul_pos < 2; mul_pos++) {
547 ir_expression *mul = op_expr[mul_pos];
548
549 if (!mul || mul->operation != ir_binop_mul)
550 continue;
551
552 /* Multiply found on one of the operands. Now check for an
553 * inner addition operation.
554 */
555 for (int inner_add_pos = 0; inner_add_pos < 2; inner_add_pos++) {
556 ir_expression *inner_add =
557 mul->operands[inner_add_pos]->as_expression();
558
559 if (!inner_add || inner_add->operation != ir_binop_add)
560 continue;
561
562 /* Inner addition found on one of the operands. Now check for
563 * one of the operands of the inner addition to be the negative
564 * of x_operand.
565 */
566 for (int neg_pos = 0; neg_pos < 2; neg_pos++) {
567 ir_expression *neg =
568 inner_add->operands[neg_pos]->as_expression();
569
570 if (!neg || neg->operation != ir_unop_neg)
571 continue;
572
573 ir_rvalue *x_operand = ir->operands[1 - mul_pos];
574
575 if (!neg->operands[0]->equals(x_operand))
576 continue;
577
578 ir_rvalue *y_operand = inner_add->operands[1 - neg_pos];
579 ir_rvalue *a_operand = mul->operands[1 - inner_add_pos];
580
581 if (!x_operand->type->is_float_16_32_64() ||
582 x_operand->type != y_operand->type ||
583 x_operand->type != a_operand->type)
584 continue;
585
586 return lrp(x_operand, y_operand, a_operand);
587 }
588 }
589 }
590
591 break;
592
593 case ir_binop_sub:
594 if (is_vec_zero(op_const[0]))
595 return neg(ir->operands[1]);
596 if (is_vec_zero(op_const[1]))
597 return ir->operands[0];
598 break;
599
600 case ir_binop_mul:
601 if (is_vec_one(op_const[0]))
602 return ir->operands[1];
603 if (is_vec_one(op_const[1]))
604 return ir->operands[0];
605
606 if (is_vec_zero(op_const[0]) || is_vec_zero(op_const[1]))
607 return ir_constant::zero(ir, ir->type);
608
609 if (is_vec_negative_one(op_const[0]))
610 return neg(ir->operands[1]);
611 if (is_vec_negative_one(op_const[1]))
612 return neg(ir->operands[0]);
613
614 if (op_expr[0] && op_expr[0]->operation == ir_unop_b2f &&
615 op_expr[1] && op_expr[1]->operation == ir_unop_b2f) {
616 return b2f(logic_and(op_expr[0]->operands[0], op_expr[1]->operands[0]));
617 }
618
619 /* Reassociate multiplication of constants so that we can do
620 * constant folding.
621 */
622 if (op_const[0] && !op_const[1])
623 reassociate_constant(ir, 0, op_const[0], op_expr[1]);
624 if (op_const[1] && !op_const[0])
625 reassociate_constant(ir, 1, op_const[1], op_expr[0]);
626
627 /* Optimizes
628 *
629 * (mul (floor (add (abs x) 0.5) (sign x)))
630 *
631 * into
632 *
633 * (trunc (add x (mul (sign x) 0.5)))
634 */
635 for (int i = 0; i < 2; i++) {
636 ir_expression *sign_expr = ir->operands[i]->as_expression();
637 ir_expression *floor_expr = ir->operands[1 - i]->as_expression();
638
639 if (!sign_expr || sign_expr->operation != ir_unop_sign ||
640 !floor_expr || floor_expr->operation != ir_unop_floor)
641 continue;
642
643 ir_expression *add_expr = floor_expr->operands[0]->as_expression();
644 if (!add_expr || add_expr->operation != ir_binop_add)
645 continue;
646
647 for (int j = 0; j < 2; j++) {
648 ir_expression *abs_expr = add_expr->operands[j]->as_expression();
649 if (!abs_expr || abs_expr->operation != ir_unop_abs)
650 continue;
651
652 ir_constant *point_five = add_expr->operands[1 - j]->as_constant();
653 if (!point_five || !point_five->is_value(0.5, 0))
654 continue;
655
656 if (abs_expr->operands[0]->equals(sign_expr->operands[0])) {
657 return trunc(add(abs_expr->operands[0],
658 mul(sign_expr, point_five)));
659 }
660 }
661 }
662 break;
663
664 case ir_binop_div:
665 if (is_vec_one(op_const[0]) && (
666 ir->type->is_float() || ir->type->is_double())) {
667 return new(mem_ctx) ir_expression(ir_unop_rcp,
668 ir->operands[1]->type,
669 ir->operands[1],
670 NULL);
671 }
672 if (is_vec_one(op_const[1]))
673 return ir->operands[0];
674 break;
675
676 case ir_binop_dot:
677 if (is_vec_zero(op_const[0]) || is_vec_zero(op_const[1]))
678 return ir_constant::zero(mem_ctx, ir->type);
679
680 for (int i = 0; i < 2; i++) {
681 if (!op_const[i])
682 continue;
683
684 unsigned components[4] = { 0 }, count = 0;
685
686 for (unsigned c = 0; c < op_const[i]->type->vector_elements; c++) {
687 if (op_const[i]->is_zero())
688 continue;
689
690 components[count] = c;
691 count++;
692 }
693
694 /* No channels had zero values; bail. */
695 if (count >= op_const[i]->type->vector_elements)
696 break;
697
698 ir_expression_operation op = count == 1 ?
699 ir_binop_mul : ir_binop_dot;
700
701 /* Swizzle both operands to remove the channels that were zero. */
702 return new(mem_ctx)
703 ir_expression(op, ir->type,
704 new(mem_ctx) ir_swizzle(ir->operands[0],
705 components, count),
706 new(mem_ctx) ir_swizzle(ir->operands[1],
707 components, count));
708 }
709 break;
710
711 case ir_binop_equal:
712 case ir_binop_nequal:
713 for (int add_pos = 0; add_pos < 2; add_pos++) {
714 ir_expression *add = op_expr[add_pos];
715
716 if (!add || add->operation != ir_binop_add)
717 continue;
718
719 ir_constant *zero = op_const[1 - add_pos];
720 if (!is_vec_zero(zero))
721 continue;
722
723 /* We are allowed to add scalars with a vector or matrix. In that
724 * case lets just exit early.
725 */
726 if (add->operands[0]->type != add->operands[1]->type)
727 continue;
728
729 /* Depending of the zero position we want to optimize
730 * (0 cmp x+y) into (-x cmp y) or (x+y cmp 0) into (x cmp -y)
731 */
732 if (add_pos == 1) {
733 return new(mem_ctx) ir_expression(ir->operation,
734 neg(add->operands[0]),
735 add->operands[1]);
736 } else {
737 return new(mem_ctx) ir_expression(ir->operation,
738 add->operands[0],
739 neg(add->operands[1]));
740 }
741 }
742 break;
743
744 case ir_binop_all_equal:
745 case ir_binop_any_nequal:
746 if (ir->operands[0]->type->is_scalar() &&
747 ir->operands[1]->type->is_scalar())
748 return new(mem_ctx) ir_expression(ir->operation == ir_binop_all_equal
749 ? ir_binop_equal : ir_binop_nequal,
750 ir->operands[0],
751 ir->operands[1]);
752 break;
753
754 case ir_binop_rshift:
755 case ir_binop_lshift:
756 /* 0 >> x == 0 */
757 if (is_vec_zero(op_const[0]))
758 return ir->operands[0];
759 /* x >> 0 == x */
760 if (is_vec_zero(op_const[1]))
761 return ir->operands[0];
762 break;
763
764 case ir_binop_logic_and:
765 if (is_vec_one(op_const[0])) {
766 return ir->operands[1];
767 } else if (is_vec_one(op_const[1])) {
768 return ir->operands[0];
769 } else if (is_vec_zero(op_const[0]) || is_vec_zero(op_const[1])) {
770 return ir_constant::zero(mem_ctx, ir->type);
771 } else if (op_expr[0] && op_expr[0]->operation == ir_unop_logic_not &&
772 op_expr[1] && op_expr[1]->operation == ir_unop_logic_not) {
773 /* De Morgan's Law:
774 * (not A) and (not B) === not (A or B)
775 */
776 return logic_not(logic_or(op_expr[0]->operands[0],
777 op_expr[1]->operands[0]));
778 } else if (ir->operands[0]->equals(ir->operands[1])) {
779 /* (a && a) == a */
780 return ir->operands[0];
781 }
782 break;
783
784 case ir_binop_logic_xor:
785 if (is_vec_zero(op_const[0])) {
786 return ir->operands[1];
787 } else if (is_vec_zero(op_const[1])) {
788 return ir->operands[0];
789 } else if (is_vec_one(op_const[0])) {
790 return logic_not(ir->operands[1]);
791 } else if (is_vec_one(op_const[1])) {
792 return logic_not(ir->operands[0]);
793 } else if (ir->operands[0]->equals(ir->operands[1])) {
794 /* (a ^^ a) == false */
795 return ir_constant::zero(mem_ctx, ir->type);
796 }
797 break;
798
799 case ir_binop_logic_or:
800 if (is_vec_zero(op_const[0])) {
801 return ir->operands[1];
802 } else if (is_vec_zero(op_const[1])) {
803 return ir->operands[0];
804 } else if (is_vec_one(op_const[0]) || is_vec_one(op_const[1])) {
805 ir_constant_data data;
806
807 for (unsigned i = 0; i < 16; i++)
808 data.b[i] = true;
809
810 return new(mem_ctx) ir_constant(ir->type, &data);
811 } else if (op_expr[0] && op_expr[0]->operation == ir_unop_logic_not &&
812 op_expr[1] && op_expr[1]->operation == ir_unop_logic_not) {
813 /* De Morgan's Law:
814 * (not A) or (not B) === not (A and B)
815 */
816 return logic_not(logic_and(op_expr[0]->operands[0],
817 op_expr[1]->operands[0]));
818 } else if (ir->operands[0]->equals(ir->operands[1])) {
819 /* (a || a) == a */
820 return ir->operands[0];
821 }
822 break;
823
824 case ir_binop_pow:
825 /* 1^x == 1 */
826 if (is_vec_one(op_const[0]))
827 return op_const[0];
828
829 /* x^1 == x */
830 if (is_vec_one(op_const[1]))
831 return ir->operands[0];
832
833 /* pow(2,x) == exp2(x) */
834 if (is_vec_two(op_const[0]))
835 return expr(ir_unop_exp2, ir->operands[1]);
836
837 if (is_vec_two(op_const[1])) {
838 ir_variable *x = new(ir) ir_variable(ir->operands[1]->type, "x",
839 ir_var_temporary);
840 base_ir->insert_before(x);
841 base_ir->insert_before(assign(x, ir->operands[0]));
842 return mul(x, x);
843 }
844
845 if (is_vec_four(op_const[1])) {
846 ir_variable *x = new(ir) ir_variable(ir->operands[1]->type, "x",
847 ir_var_temporary);
848 base_ir->insert_before(x);
849 base_ir->insert_before(assign(x, ir->operands[0]));
850
851 ir_variable *squared = new(ir) ir_variable(ir->operands[1]->type,
852 "squared",
853 ir_var_temporary);
854 base_ir->insert_before(squared);
855 base_ir->insert_before(assign(squared, mul(x, x)));
856 return mul(squared, squared);
857 }
858
859 break;
860
861 case ir_binop_min:
862 case ir_binop_max:
863 if (!ir->type->is_float())
864 break;
865
866 /* Replace min(max) operations and its commutative combinations with
867 * a saturate operation
868 */
869 for (int op = 0; op < 2; op++) {
870 ir_expression *inner_expr = op_expr[op];
871 ir_constant *outer_const = op_const[1 - op];
872 ir_expression_operation op_cond = (ir->operation == ir_binop_max) ?
873 ir_binop_min : ir_binop_max;
874
875 if (!inner_expr || !outer_const || (inner_expr->operation != op_cond))
876 continue;
877
878 /* One of these has to be a constant */
879 if (!inner_expr->operands[0]->as_constant() &&
880 !inner_expr->operands[1]->as_constant())
881 break;
882
883 /* Found a min(max) combination. Now try to see if its operands
884 * meet our conditions that we can do just a single saturate operation
885 */
886 for (int minmax_op = 0; minmax_op < 2; minmax_op++) {
887 ir_rvalue *x = inner_expr->operands[minmax_op];
888 ir_rvalue *y = inner_expr->operands[1 - minmax_op];
889
890 ir_constant *inner_const = y->as_constant();
891 if (!inner_const)
892 continue;
893
894 /* min(max(x, 0.0), 1.0) is sat(x) */
895 if (ir->operation == ir_binop_min &&
896 inner_const->is_zero() &&
897 outer_const->is_one())
898 return saturate(x);
899
900 /* max(min(x, 1.0), 0.0) is sat(x) */
901 if (ir->operation == ir_binop_max &&
902 inner_const->is_one() &&
903 outer_const->is_zero())
904 return saturate(x);
905
906 /* min(max(x, 0.0), b) where b < 1.0 is sat(min(x, b)) */
907 if (ir->operation == ir_binop_min &&
908 inner_const->is_zero() &&
909 is_less_than_one(outer_const))
910 return saturate(expr(ir_binop_min, x, outer_const));
911
912 /* max(min(x, b), 0.0) where b < 1.0 is sat(min(x, b)) */
913 if (ir->operation == ir_binop_max &&
914 is_less_than_one(inner_const) &&
915 outer_const->is_zero())
916 return saturate(expr(ir_binop_min, x, inner_const));
917
918 /* max(min(x, 1.0), b) where b > 0.0 is sat(max(x, b)) */
919 if (ir->operation == ir_binop_max &&
920 inner_const->is_one() &&
921 is_greater_than_zero(outer_const))
922 return saturate(expr(ir_binop_max, x, outer_const));
923
924 /* min(max(x, b), 1.0) where b > 0.0 is sat(max(x, b)) */
925 if (ir->operation == ir_binop_min &&
926 is_greater_than_zero(inner_const) &&
927 outer_const->is_one())
928 return saturate(expr(ir_binop_max, x, inner_const));
929 }
930 }
931
932 break;
933
934 case ir_unop_rcp:
935 if (op_expr[0] && op_expr[0]->operation == ir_unop_rcp)
936 return op_expr[0]->operands[0];
937
938 if (op_expr[0] && (op_expr[0]->operation == ir_unop_exp2 ||
939 op_expr[0]->operation == ir_unop_exp)) {
940 return new(mem_ctx) ir_expression(op_expr[0]->operation, ir->type,
941 neg(op_expr[0]->operands[0]));
942 }
943
944 if (op_expr[0] && op_expr[0]->operation == ir_unop_rsq)
945 return sqrt(op_expr[0]->operands[0]);
946
947 /* As far as we know, all backends are OK with rsq. */
948 if (op_expr[0] && op_expr[0]->operation == ir_unop_sqrt) {
949 return rsq(op_expr[0]->operands[0]);
950 }
951
952 break;
953
954 case ir_triop_fma:
955 /* Operands are op0 * op1 + op2. */
956 if (is_vec_zero(op_const[0]) || is_vec_zero(op_const[1])) {
957 return ir->operands[2];
958 } else if (is_vec_zero(op_const[2])) {
959 return mul(ir->operands[0], ir->operands[1]);
960 } else if (is_vec_one(op_const[0])) {
961 return add(ir->operands[1], ir->operands[2]);
962 } else if (is_vec_one(op_const[1])) {
963 return add(ir->operands[0], ir->operands[2]);
964 }
965 break;
966
967 case ir_triop_lrp:
968 /* Operands are (x, y, a). */
969 if (is_vec_zero(op_const[2])) {
970 return ir->operands[0];
971 } else if (is_vec_one(op_const[2])) {
972 return ir->operands[1];
973 } else if (ir->operands[0]->equals(ir->operands[1])) {
974 return ir->operands[0];
975 } else if (is_vec_zero(op_const[0])) {
976 return mul(ir->operands[1], ir->operands[2]);
977 } else if (is_vec_zero(op_const[1])) {
978 unsigned op2_components = ir->operands[2]->type->vector_elements;
979 ir_constant *one;
980
981 switch (ir->type->base_type) {
982 case GLSL_TYPE_FLOAT16:
983 one = new(mem_ctx) ir_constant(float16_t::one(), op2_components);
984 break;
985 case GLSL_TYPE_FLOAT:
986 one = new(mem_ctx) ir_constant(1.0f, op2_components);
987 break;
988 case GLSL_TYPE_DOUBLE:
989 one = new(mem_ctx) ir_constant(1.0, op2_components);
990 break;
991 default:
992 one = NULL;
993 unreachable("unexpected type");
994 }
995
996 return mul(ir->operands[0], add(one, neg(ir->operands[2])));
997 }
998 break;
999
1000 case ir_triop_csel:
1001 if (is_vec_one(op_const[0]))
1002 return ir->operands[1];
1003 if (is_vec_zero(op_const[0]))
1004 return ir->operands[2];
1005 break;
1006
1007 /* Remove interpolateAt* instructions for demoted inputs. They are
1008 * assigned a constant expression to facilitate this.
1009 */
1010 case ir_unop_interpolate_at_centroid:
1011 case ir_binop_interpolate_at_offset:
1012 case ir_binop_interpolate_at_sample:
1013 if (op_const[0])
1014 return ir->operands[0];
1015 break;
1016
1017 default:
1018 break;
1019 }
1020
1021 return ir;
1022 }
1023
1024 void
handle_rvalue(ir_rvalue ** rvalue)1025 ir_algebraic_visitor::handle_rvalue(ir_rvalue **rvalue)
1026 {
1027 if (!*rvalue)
1028 return;
1029
1030 ir_expression *expr = (*rvalue)->as_expression();
1031 if (!expr || expr->operation == ir_quadop_vector)
1032 return;
1033
1034 ir_rvalue *new_rvalue = handle_expression(expr);
1035 if (new_rvalue == *rvalue)
1036 return;
1037
1038 /* If the expr used to be some vec OP scalar returning a vector, and the
1039 * optimization gave us back a scalar, we still need to turn it into a
1040 * vector.
1041 */
1042 *rvalue = swizzle_if_required(expr, new_rvalue);
1043
1044 this->progress = true;
1045 }
1046
1047 bool
do_algebraic(exec_list * instructions,bool native_integers,const struct gl_shader_compiler_options * options)1048 do_algebraic(exec_list *instructions, bool native_integers,
1049 const struct gl_shader_compiler_options *options)
1050 {
1051 ir_algebraic_visitor v(native_integers, options);
1052
1053 visit_list_elements(&v, instructions);
1054
1055 return v.progress;
1056 }
1057