/external/llvm-project/llvm/test/CodeGen/X86/ |
D | pr46877.ll | 206 %81 = fsub reassoc nsz contract float %0, %1 207 %82 = fmul reassoc nsz contract float %1, %2 208 %83 = fmul reassoc nsz contract float %3, %82 209 %84 = fsub reassoc nsz contract float %0, %83 210 %85 = fmul reassoc nsz contract float %84, %4 211 %86 = fmul reassoc nsz contract float %81, %5 212 %87 = fsub reassoc nsz contract float %0, %86 213 %88 = fmul reassoc nsz contract float %87, %85 214 %89 = fmul reassoc nsz contract float %81, %6 215 %90 = fmul reassoc nsz contract float %89, %7 [all …]
|
D | fp-fast.ll | 8 %t1 = fadd nnan reassoc nsz float %a, %a 9 %r = fadd nnan reassoc nsz float %t1, %t1 18 %t1 = fmul nnan reassoc nsz float 4.0, %a 19 %t2 = fadd nnan reassoc nsz float %a, %a 20 %r = fadd nnan reassoc nsz float %t1, %t2 29 %t1 = fmul nnan reassoc nsz float %a, 4.0 30 %t2 = fadd nnan reassoc nsz float %a, %a 31 %r = fadd nnan reassoc nsz float %t1, %t2 40 %t1 = fadd nnan reassoc nsz float %a, %a 41 %t2 = fmul nnan reassoc nsz float 4.0, %a [all …]
|
D | sqrt-fastmath-mir.ll | 99 …; CHECK: [[VRSQRTSSr:%[0-9]+]]:fr32 = nnan ninf nsz arcp contract afn reassoc VRSQRTSSr killed [… 100 …; CHECK: %3:fr32 = nnan ninf nsz arcp contract afn reassoc nofpexcept VMULSSrr [[COPY]], [[VRSQR… 102 …; CHECK: %5:fr32 = nnan ninf nsz arcp contract afn reassoc nofpexcept VFMADD213SSr [[VRSQRTSSr]]… 104 …; CHECK: %7:fr32 = nnan ninf nsz arcp contract afn reassoc nofpexcept VMULSSrr [[VRSQRTSSr]], [[… 105 …; CHECK: %8:fr32 = nnan ninf nsz arcp contract afn reassoc nofpexcept VMULSSrr killed %7, killed… 106 …; CHECK: %9:fr32 = nnan ninf nsz arcp contract afn reassoc nofpexcept VMULSSrr [[COPY]], %8, imp… 107 …; CHECK: %10:fr32 = nnan ninf nsz arcp contract afn reassoc nofpexcept VFMADD213SSr %8, killed %… 108 …; CHECK: %11:fr32 = nnan ninf nsz arcp contract afn reassoc nofpexcept VMULSSrr %8, [[VMOVSSrm_a… 109 …; CHECK: %12:fr32 = nnan ninf nsz arcp contract afn reassoc nofpexcept VMULSSrr killed %11, kill… 123 …; CHECK: [[VRSQRTSSr:%[0-9]+]]:fr32 = nnan ninf nsz arcp contract afn reassoc VRSQRTSSr killed [… [all …]
|
D | pow.75.ll | 31 ; CHECK: Combining: {{.*}}: v4f32 = fpow nnan ninf nsz arcp contract afn reassoc [[X:t[0-9]+]], {{.… 32 …T: Creating new node: [[SQRT:t[0-9]+]]: v4f32 = fsqrt nnan ninf nsz arcp contract afn reassoc [[X]] 33 …ting new node: [[SQRTSQRT:t[0-9]+]]: v4f32 = fsqrt nnan ninf nsz arcp contract afn reassoc [[SQRT]] 34 …ng new node: [[R:t[0-9]+]]: v4f32 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQRTS… 35 ; CHECK-NEXT: ... into: [[R]]: v4f32 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQ… 41 ; CHECK: Combining: {{.*}}: v2f64 = fpow nnan ninf nsz arcp contract afn reassoc [[X:t[0-9]+]], {{.… 42 …T: Creating new node: [[SQRT:t[0-9]+]]: v2f64 = fsqrt nnan ninf nsz arcp contract afn reassoc [[X]] 43 …ting new node: [[SQRTSQRT:t[0-9]+]]: v2f64 = fsqrt nnan ninf nsz arcp contract afn reassoc [[SQRT]] 44 …ng new node: [[R:t[0-9]+]]: v2f64 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQRTS… 45 ; CHECK-NEXT: ... into: [[R]]: v2f64 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQ…
|
/external/llvm-project/llvm/test/CodeGen/PowerPC/ |
D | machine-combiner.ll | 17 %t0 = fadd reassoc nsz float %x0, %x1 18 %t1 = fadd reassoc nsz float %t0, %x2 19 %t2 = fadd reassoc nsz float %t1, %x3 31 %t0 = fadd reassoc nsz float %x0, %x1 32 %t1 = fadd reassoc nsz float %x2, %t0 33 %t2 = fadd reassoc nsz float %t1, %x3 45 %t0 = fadd reassoc nsz float %x0, %x1 46 %t1 = fadd reassoc nsz float %t0, %x2 47 %t2 = fadd reassoc nsz float %x3, %t1 59 %t0 = fadd reassoc nsz float %x0, %x1 [all …]
|
D | fma-precision.ll | 14 %mul = fmul reassoc double %b, %a 15 %mul1 = fmul reassoc double %d, %c 16 %sub = fsub reassoc nsz double %mul, %mul1 17 %mul3 = fmul reassoc double %mul, %sub 31 %mul = fmul reassoc double %b, %a 32 %mul1 = fmul reassoc double %d, %c 33 %sub = fsub reassoc double %mul, %mul1 34 %mul3 = fmul reassoc double %mul1, %sub 47 %mul = fmul reassoc double %b, %a 48 %mul1 = fmul reassoc double %d, %c [all …]
|
D | fma-negate.ll | 20 %0 = fmul contract reassoc double %b, %c 21 %1 = fsub contract reassoc double %a, %0 39 %0 = fmul contract reassoc double %a, %b 40 %1 = fmul contract reassoc double %c, %d 41 %2 = fsub contract reassoc double %0, %1 59 %0 = fsub contract reassoc double -0.0, %a 60 %1 = call contract reassoc double @llvm.fma.f64(double %0, double %b, 78 %0 = fmul contract reassoc float %b, %c 79 %1 = fsub contract reassoc float %a, %0 97 %0 = fmul contract reassoc float %a, %b [all …]
|
D | fma-assoc.ll | 290 %F = fmul reassoc double %A, %B ; <double> [#uses=1] 291 %G = fmul reassoc double %C, %D ; <double> [#uses=1] 292 %H = fadd reassoc double %F, %G ; <double> [#uses=1] 293 %I = fadd reassoc double %H, %E ; <double> [#uses=1] 311 %F = fmul reassoc double %A, %B ; <double> [#uses=1] 312 %G = fmul reassoc double %C, %D ; <double> [#uses=1] 313 %H = fadd reassoc double %F, %G ; <double> [#uses=1] 314 %I = fadd reassoc double %E, %H ; <double> [#uses=1] 332 %F = fmul reassoc double %A, %B ; <double> [#uses=1] 333 %G = fmul reassoc double %C, %D ; <double> [#uses=1] [all …]
|
/external/llvm-project/llvm/test/Transforms/Reassociate/ |
D | fast-basictest.ll | 16 ; Both 'reassoc' and 'nsz' are required. 19 ; CHECK-NEXT: [[TMP1:%.*]] = fneg reassoc nsz float [[ARG:%.*]] 22 %t1 = fsub reassoc nsz float -1.200000e+01, %arg 23 %t2 = fadd reassoc nsz float %t1, 1.200000e+01 27 ; Verify the fold is not done with only 'reassoc' ('nsz' is required). 30 ; CHECK-NEXT: [[T1:%.*]] = fsub reassoc float -1.200000e+01, [[ARG:%.*]] 31 ; CHECK-NEXT: [[T2:%.*]] = fadd reassoc float [[T1]], 1.200000e+01 34 %t1 = fsub reassoc float -1.200000e+01, %arg 35 %t2 = fadd reassoc float %t1, 1.200000e+01 65 ; CHECK-NEXT: [[REG115:%.*]] = fadd reassoc float [[REG109:%.*]], -3.000000e+01 [all …]
|
D | fast-SubReassociate.ll | 33 ; Both 'reassoc' and 'nsz' are required. 36 ; CHECK-NEXT: [[TMP1:%.*]] = fsub reassoc nsz float [[A:%.*]], [[B:%.*]] 39 %W = fadd reassoc nsz float %B, 5.000000e+00 40 %X = fadd reassoc nsz float %A, -7.000000e+00 41 %Y = fsub reassoc nsz float %X, %W 42 %Z = fadd reassoc nsz float %Y, 1.200000e+01 46 ; Verify the fold is not done with only 'reassoc' ('nsz' is required). 49 ; CHECK-NEXT: [[W:%.*]] = fadd reassoc float [[B:%.*]], 5.000000e+00 50 ; CHECK-NEXT: [[X:%.*]] = fadd reassoc float [[A:%.*]], -7.000000e+00 51 ; CHECK-NEXT: [[Y:%.*]] = fsub reassoc float [[X]], [[W]] [all …]
|
D | fast-ReassociateVector.ll | 22 ; CHECK-NEXT: [[MUL:%.*]] = fmul reassoc <4 x float> [[A:%.*]], [[C:%.*]] 23 ; CHECK-NEXT: [[MUL1:%.*]] = fmul reassoc <4 x float> [[B:%.*]], [[C]] 24 ; CHECK-NEXT: [[ADD:%.*]] = fadd reassoc <4 x float> [[MUL]], [[MUL1]] 27 %mul = fmul reassoc <4 x float> %a, %c 28 %mul1 = fmul reassoc <4 x float> %b, %c 29 %add = fadd reassoc <4 x float> %mul, %mul1 54 ; CHECK-NEXT: [[T0:%.*]] = fmul reassoc <2 x float> [[A:%.*]], [[B:%.*]] 55 ; CHECK-NEXT: [[T1:%.*]] = fmul reassoc <2 x float> [[A]], [[T0]] 56 ; CHECK-NEXT: [[T2:%.*]] = fmul reassoc <2 x float> [[A]], [[C:%.*]] 57 ; CHECK-NEXT: [[T3:%.*]] = fmul reassoc <2 x float> [[A]], [[T2]] [all …]
|
D | fast-MissedTree.ll | 15 ; Both 'reassoc' and 'nsz' are required. 18 ; CHECK-NEXT: [[Z:%.*]] = fadd reassoc nsz float %A, %B 21 %W = fadd reassoc nsz float %B, -5.0 22 %Y = fadd reassoc nsz float %A, 5.0 23 %Z = fadd reassoc nsz float %W, %Y 27 ; Verify the fold is not done with only 'reassoc' ('nsz' is required). 30 ; CHECK-NEXT: [[W:%.*]] = fadd reassoc float %B, -5.000000e+00 31 ; CHECK-NEXT: [[Y:%.*]] = fadd reassoc float %A, 5.000000e+00 32 ; CHECK-NEXT: [[Z:%.*]] = fadd reassoc float [[Y]], [[W]] 35 %W = fadd reassoc float %B, -5.0 [all …]
|
D | fast-AgressiveSubMove.ll | 30 ; CHECK-NEXT: [[X:%.*]] = fadd reassoc float %A, 1.000000e+00 31 ; CHECK-NEXT: [[Y:%.*]] = fadd reassoc float %A, 1.000000e+00 32 ; CHECK-NEXT: [[R:%.*]] = fsub reassoc float [[X]], [[Y]] 35 %X = fadd reassoc float 1.000000e+00, %A 36 %Y = fadd reassoc float 1.000000e+00, %A 37 %r = fsub reassoc float %X, %Y
|
/external/llvm-project/llvm/test/Transforms/InstCombine/ |
D | fdiv-cos-sin.ll | 20 ; CHECK-NEXT: [[TMP2:%.*]] = call reassoc double @llvm.sin.f64(double [[A]]) 25 %2 = call reassoc double @llvm.sin.f64(double %a) 32 ; CHECK-NEXT: [[TAN:%.*]] = call reassoc double @tan(double [[A:%.*]]) #1 33 ; CHECK-NEXT: [[TMP1:%.*]] = fdiv reassoc double 1.000000e+00, [[TAN]] 38 %div = fdiv reassoc double %1, %2 44 ; CHECK-NEXT: [[TAN:%.*]] = call reassoc double @tan(double [[A:%.*]]) #1 45 ; CHECK-NEXT: [[TMP1:%.*]] = fdiv reassoc double 1.000000e+00, [[TAN]] 48 %1 = call reassoc double @llvm.cos.f64(double %a) 50 %div = fdiv reassoc double %1, %2 56 ; CHECK-NEXT: [[TMP1:%.*]] = call reassoc double @llvm.cos.f64(double [[A:%.*]]) [all …]
|
D | fdiv-sin-cos.ll | 20 ; CHECK-NEXT: [[TMP2:%.*]] = call reassoc double @llvm.cos.f64(double [[A]]) 25 %2 = call reassoc double @llvm.cos.f64(double %a) 32 ; CHECK-NEXT: [[TAN:%.*]] = call reassoc double @tan(double [[A:%.*]]) #1 37 %div = fdiv reassoc double %1, %2 43 ; CHECK-NEXT: [[TAN:%.*]] = call reassoc double @tan(double [[A:%.*]]) #1 46 %1 = call reassoc double @llvm.sin.f64(double %a) 48 %div = fdiv reassoc double %1, %2 54 ; CHECK-NEXT: [[TMP1:%.*]] = call reassoc double @llvm.sin.f64(double [[A:%.*]]) 55 ; CHECK-NEXT: [[TMP2:%.*]] = call reassoc double @llvm.cos.f64(double [[A]]) 56 ; CHECK-NEXT: [[DIV:%.*]] = fdiv reassoc double [[TMP1]], [[TMP2]] [all …]
|
D | fmul-exp2.ll | 7 ; exp2(a) * exp2(b) no reassoc flags 21 ; exp2(a) * exp2(b) reassoc, multiple uses 25 ; CHECK-NEXT: [[TMP:%.*]] = fadd reassoc double [[A:%.*]], [[B]] 26 ; CHECK-NEXT: [[TMP2:%.*]] = call reassoc double @llvm.exp2.f64(double [[TMP]]) 32 %mul = fmul reassoc double %tmp, %tmp1 37 ; exp2(a) * exp2(b) reassoc, both with multiple uses 42 ; CHECK-NEXT: [[MUL:%.*]] = fmul reassoc double [[TMP]], [[TMP1]] 49 %mul = fmul reassoc double %tmp, %tmp1 55 ; exp2(a) * exp2(b) => exp2(a+b) with reassoc 58 ; CHECK-NEXT: [[TMP:%.*]] = fadd reassoc double [[A:%.*]], [[B:%.*]] [all …]
|
D | fmul-exp.ll | 7 ; exp(a) * exp(b) no reassoc flags 21 ; exp(a) * exp(b) reassoc, multiple uses 25 ; CHECK-NEXT: [[TMP:%.*]] = fadd reassoc double [[A:%.*]], [[B]] 26 ; CHECK-NEXT: [[TMP2:%.*]] = call reassoc double @llvm.exp.f64(double [[TMP]]) 32 %mul = fmul reassoc double %tmp, %tmp1 37 ; exp(a) * exp(b) reassoc, both with multiple uses 42 ; CHECK-NEXT: [[MUL:%.*]] = fmul reassoc double [[TMP]], [[TMP1]] 49 %mul = fmul reassoc double %tmp, %tmp1 55 ; exp(a) * exp(b) => exp(a+b) with reassoc 58 ; CHECK-NEXT: [[TMP:%.*]] = fadd reassoc double [[A:%.*]], [[B:%.*]] [all …]
|
D | fast-math.ll | 53 ; Check again with 'reassoc' and 'nsz' ('nsz' not technically required). 56 ; CHECK-NEXT: [[TMP1:%.*]] = fmul reassoc nsz double [[F1:%.*]], 6.000000e+00 59 %t1 = fmul reassoc nsz double 5.000000e+00, %f1 60 %t2 = fadd reassoc nsz double %f1, %t1 67 ; CHECK-NEXT: [[T1:%.*]] = fmul reassoc double [[F1:%.*]], 5.000000e+00 68 ; CHECK-NEXT: [[T2:%.*]] = fadd reassoc double [[T1]], [[F1]] 71 %t1 = fmul reassoc double 5.000000e+00, %f1 72 %t2 = fadd reassoc double %f1, %t1 89 ; Check again with 'reassoc' and 'nsz' ('nsz' not technically required). 92 ; CHECK-NEXT: [[TMP1:%.*]] = fadd reassoc nsz float [[F1:%.*]], [[F2:%.*]] [all …]
|
D | 2006-10-26-VectorReassoc.ll | 27 ; Verify this folds with 'reassoc' and 'nsz' ('nsz' not technically required) 30 ; CHECK-NEXT: [[TMP1:%.*]] = fmul reassoc nsz <4 x float> [[V:%.*]], <float 1.000000e+00, float… 32 …%Y = fmul reassoc nsz <4 x float> %V, < float 1.000000e+00, float 2.000000e+00, float 3.000000e+00… 33 …%Z = fmul reassoc nsz <4 x float> %Y, < float 1.000000e+00, float 2.000000e+05, float -3.000000e+0… 41 ; CHECK-NEXT: [[TMP1:%.*]] = fmul reassoc <4 x float> [[V:%.*]], <float 1.000000e+00, float 2.0… 42 ; CHECK-NEXT: [[TMP2:%.*]] = fmul reassoc <4 x float> [[TMP1]], <float 1.000000e+00, float 2.00… 44 …%Y = fmul reassoc <4 x float> %V, < float 1.000000e+00, float 2.000000e+00, float 3.000000e+00, fl… 45 …%Z = fmul reassoc <4 x float> %Y, < float 1.000000e+00, float 2.000000e+05, float -3.000000e+00, f… 73 ; Verify this folds with 'reassoc' and 'nsz' ('nsz' not technically required) 76 ; CHECK-NEXT: [[TMP1:%.*]] = fadd reassoc nsz <4 x float> [[V:%.*]], <float 2.000000e+00, float… [all …]
|
D | fadd-fsub-factor.ll | 18 ; CHECK-NEXT: [[TMP1:%.*]] = fadd reassoc nsz float [[X:%.*]], [[Y:%.*]] 19 ; CHECK-NEXT: [[R:%.*]] = fmul reassoc nsz float [[TMP1]], [[Z:%.*]] 24 %r = fadd reassoc nsz float %t1, %t2 32 ; CHECK-NEXT: [[TMP1:%.*]] = fadd reassoc nsz <2 x float> [[X:%.*]], [[Y:%.*]] 33 ; CHECK-NEXT: [[R:%.*]] = fmul reassoc nsz <2 x float> [[TMP1]], [[Z:%.*]] 38 %r = fadd reassoc nsz <2 x float> %t1, %t2 46 ; CHECK-NEXT: [[TMP1:%.*]] = fadd reassoc ninf nsz <2 x float> [[X:%.*]], [[Y:%.*]] 47 ; CHECK-NEXT: [[R:%.*]] = fmul reassoc ninf nsz <2 x float> [[TMP1]], [[Z:%.*]] 52 %r = fadd reassoc nsz ninf <2 x float> %t1, %t2 60 ; CHECK-NEXT: [[TMP1:%.*]] = fadd reassoc nnan nsz double [[X:%.*]], [[Y:%.*]] [all …]
|
D | fmul-sqrt.ll | 44 ; CHECK-NEXT: [[TMP1:%.*]] = fmul reassoc nnan double [[A:%.*]], [[B:%.*]] 45 ; CHECK-NEXT: [[TMP2:%.*]] = call reassoc nnan double @llvm.sqrt.f64(double [[TMP1]]) 50 %mul = fmul reassoc nnan double %1, %2 61 ; CHECK-NEXT: [[MUL:%.*]] = fmul reassoc double [[TMP1]], [[TMP2]] 66 %mul = fmul reassoc double %1, %2 71 ; 'reassoc nnan' on the fmuls is all that is required, but check propagation of other FMF. 75 ; CHECK-NEXT: [[TMP1:%.*]] = fmul reassoc nnan arcp double [[A:%.*]], [[B:%.*]] 76 ; CHECK-NEXT: [[TMP2:%.*]] = fmul reassoc nnan double [[TMP1]], [[C:%.*]] 77 ; CHECK-NEXT: [[TMP3:%.*]] = fmul reassoc nnan ninf double [[TMP2]], [[D:%.*]] 78 ; CHECK-NEXT: [[TMP4:%.*]] = call reassoc nnan ninf double @llvm.sqrt.f64(double [[TMP3]]) [all …]
|
/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | pow.75.ll | 32 ; CHECK: Combining: [[FIRST]]: f32 = fpow nnan ninf nsz arcp contract afn reassoc [[X:t[0-9]+]], Co… 33 …EXT: Creating new node: [[SQRT:t[0-9]+]]: f32 = fsqrt nnan ninf nsz arcp contract afn reassoc [[X]] 34 …eating new node: [[SQRTSQRT:t[0-9]+]]: f32 = fsqrt nnan ninf nsz arcp contract afn reassoc [[SQRT]] 35 ; CHECK-NEXT: Creating new node: [[R:t[0-9]+]]: f32 = fmul nnan ninf nsz arcp contract afn reassoc … 36 ; CHECK-NEXT: ... into: [[R]]: f32 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQRT… 37 ; CHECK: Combining: [[SECOND]]: f32 = fpow nnan ninf nsz arcp contract afn reassoc [[X:t[0-9]+]], C… 38 …EXT: Creating new node: [[SQRT:t[0-9]+]]: f32 = fsqrt nnan ninf nsz arcp contract afn reassoc [[X]] 39 …eating new node: [[SQRTSQRT:t[0-9]+]]: f32 = fsqrt nnan ninf nsz arcp contract afn reassoc [[SQRT]] 40 ; CHECK-NEXT: Creating new node: [[R:t[0-9]+]]: f32 = fmul nnan ninf nsz arcp contract afn reassoc … 41 ; CHECK-NEXT: ... into: [[R]]: f32 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQRT… [all …]
|
/external/llvm-project/llvm/test/Transforms/InstSimplify/ |
D | log-exp-intrinsic.ll | 14 %2 = call reassoc double @llvm.log.f64(double %1) 20 ; CHECK-NEXT: [[TMP1:%.*]] = call reassoc double @llvm.exp.f64(double [[A:%.*]]) 24 %1 = call reassoc double @llvm.exp.f64(double %a) 49 %2 = call reassoc double @llvm.log.f64(double %1) 51 %4 = call reassoc double @llvm.log.f64(double %3) 60 %2 = call reassoc double @llvm.log2.f64(double %1) 66 ; CHECK-NEXT: [[TMP1:%.*]] = call reassoc double @llvm.exp2.f64(double [[A:%.*]]) 70 %1 = call reassoc double @llvm.exp2.f64(double %a) 95 %2 = call reassoc double @llvm.log2.f64(double %1) 97 %4 = call reassoc double @llvm.log2.f64(double %3) [all …]
|
/external/llvm-project/llvm/test/CodeGen/MIR/X86/ |
D | fastmath.mir | 25 ; CHECK: %7:fr32 = reassoc VMULSSrr %6, %6, implicit $mxcsr 26 %7:fr32 = reassoc VMULSSrr %6, %6, implicit $mxcsr 27 ; CHECK: %8:fr32 = nsz arcp contract afn reassoc VMULSSrr %7, %7, implicit $mxcsr 28 %8:fr32 = nsz arcp contract afn reassoc VMULSSrr %7, %7, implicit $mxcsr 29 ; CHECK: %9:fr32 = contract afn reassoc VMULSSrr %8, %8, implicit $mxcsr 30 %9:fr32 = contract afn reassoc VMULSSrr %8, %8, implicit $mxcsr
|
/external/llvm-project/llvm/test/CodeGen/AArch64/ |
D | pow.75.ll | 31 ; CHECK: Combining: {{.*}}: v4f32 = fpow nnan ninf nsz arcp contract afn reassoc [[X:t[0-9]+]], {{.… 32 …T: Creating new node: [[SQRT:t[0-9]+]]: v4f32 = fsqrt nnan ninf nsz arcp contract afn reassoc [[X]] 33 …ting new node: [[SQRTSQRT:t[0-9]+]]: v4f32 = fsqrt nnan ninf nsz arcp contract afn reassoc [[SQRT]] 34 …ng new node: [[R:t[0-9]+]]: v4f32 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQRTS… 35 ; CHECK-NEXT: ... into: [[R]]: v4f32 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQ… 41 ; CHECK: Combining: {{.*}}: v2f64 = fpow nnan ninf nsz arcp contract afn reassoc [[X:t[0-9]+]], {{.… 42 …T: Creating new node: [[SQRT:t[0-9]+]]: v2f64 = fsqrt nnan ninf nsz arcp contract afn reassoc [[X]] 43 …ting new node: [[SQRTSQRT:t[0-9]+]]: v2f64 = fsqrt nnan ninf nsz arcp contract afn reassoc [[SQRT]] 44 …ng new node: [[R:t[0-9]+]]: v2f64 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQRTS… 45 ; CHECK-NEXT: ... into: [[R]]: v2f64 = fmul nnan ninf nsz arcp contract afn reassoc [[SQRT]], [[SQ…
|