/external/llvm/test/Transforms/Reassociate/ |
D | fast-basictest.ll | 6 ; CHECK-NEXT: fsub fast float -0.000000e+00, %arg 9 %tmp1 = fsub fast float -1.200000e+01, %arg 10 %tmp2 = fadd fast float %tmp1, 1.200000e+01 29 ; CHECK-NEXT: %reg117 = fadd fast float %reg109, %reg1111 32 %reg115 = fadd fast float %reg109, -3.000000e+01 33 %reg116 = fadd fast float %reg115, %reg1111 34 %reg117 = fadd fast float %reg116, 3.000000e+01 46 ; CHECK: fadd fast float 47 ; CHECK: fadd fast float 48 ; CHECK-NOT: fadd fast float [all …]
|
D | fast-mightymul.ll | 5 %t0 = fmul fast float %x, %x 6 %t1 = fmul fast float %t0, %t0 7 %t2 = fmul fast float %t1, %t1 8 %t3 = fmul fast float %t2, %t2 9 %t4 = fmul fast float %t3, %t3 10 %t5 = fmul fast float %t4, %t4 11 %t6 = fmul fast float %t5, %t5 12 %t7 = fmul fast float %t6, %t6 13 %t8 = fmul fast float %t7, %t7 14 %t9 = fmul fast float %t8, %t8 [all …]
|
D | fast-fp-commute.ll | 7 ; CHECK: fmul fast float %y, %x 8 ; CHECK: fmul fast float %y, %x 9 ; CHECK: fsub fast float %1, %2 13 %1 = fmul fast float %x, %y 14 %2 = fmul fast float %y, %x 15 %3 = fsub fast float %1, %2 23 ; CHECK-NEXT: fmul fast float %y, %x 24 ; CHECK-NEXT: fmul fast float %y, %x 25 ; CHECK-NEXT: fsub fast float %1, %2 28 %1 = fmul fast float %x, %y [all …]
|
D | fast-ReassociateVector.ll | 6 ; CHECK-NEXT: %tmp = fadd fast <4 x float> %b, %a 7 ; CHECK-NEXT: %tmp1 = fmul fast <4 x float> %tmp, %c 10 %mul = fmul fast <4 x float> %a, %c 11 %mul1 = fmul fast <4 x float> %b, %c 12 %add = fadd fast <4 x float> %mul, %mul1 19 ; CHECK-NEXT: [[TMP1:%tmp.*]] = fadd fast <2 x float> %c, %b 20 ; CHECK-NEXT: [[TMP2:%tmp.*]] = fmul fast <2 x float> %a, %a 21 ; CHECK-NEXT: fmul fast <2 x float> [[TMP2]], [[TMP1]] 24 %t0 = fmul fast <2 x float> %a, %b 25 %t1 = fmul fast <2 x float> %a, %t0 [all …]
|
D | reassoc-intermediate-fnegs.ll | 3 ; CHECK: [[TMP1:%.*]] = fsub fast half 0xH8000, %a 4 ; CHECK: [[TMP2:%.*]] = fadd fast half %b, [[TMP1]] 5 ; CHECK: fmul fast half [[TMP2]], 0xH4500 9 %tmp1 = fmul fast half %b, 0xH4200 ; 3*b 10 %tmp2 = fmul fast half %a, 0xH4500 ; 5*a 11 %tmp3 = fmul fast half %b, 0xH4000 ; 2*b 12 %tmp4 = fsub fast half %tmp2, %tmp1 ; 5 * a - 3 * b 13 %tmp5 = fsub fast half %tmp3, %tmp4 ; 2 * b - ( 5 * a - 3 * b) 18 ; CHECK: [[TMP1:%tmp.*]] = fmul fast half %a, 0xH4500 19 ; CHECK: [[TMP2:%tmp.*]] = fmul fast half %b, 0xH3C00 [all …]
|
D | fast-multistep.ll | 6 ; CHECK-NEXT: [[TMP1:%tmp.*]] = fadd fast float %c, %b 7 ; CHECK-NEXT: [[TMP2:%tmp.*]] = fmul fast float %a, %a 8 ; CHECK-NEXT: fmul fast float [[TMP2]], [[TMP1]] 11 %t0 = fmul fast float %a, %b 12 %t1 = fmul fast float %a, %t0 ; a*(a*b) 13 %t2 = fmul fast float %a, %c 14 %t3 = fmul fast float %a, %t2 ; a*(a*c) 15 %t4 = fadd fast float %t1, %t3 22 ; CHECK-NEXT: fadd fast float %c, %b 23 ; CHECK-NEXT: fmul fast float %tmp, %a [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/Reassociate/ |
D | fast-mightymul.ll | 5 %t0 = fmul fast float %x, %x 6 %t1 = fmul fast float %t0, %t0 7 %t2 = fmul fast float %t1, %t1 8 %t3 = fmul fast float %t2, %t2 9 %t4 = fmul fast float %t3, %t3 10 %t5 = fmul fast float %t4, %t4 11 %t6 = fmul fast float %t5, %t5 12 %t7 = fmul fast float %t6, %t6 13 %t8 = fmul fast float %t7, %t7 14 %t9 = fmul fast float %t8, %t8 [all …]
|
D | fast-fp-commute.ll | 7 ; CHECK-NEXT: [[TMP1:%.*]] = fmul fast float %y, %x 8 ; CHECK-NEXT: [[TMP2:%.*]] = fmul fast float %y, %x 9 ; CHECK-NEXT: [[TMP3:%.*]] = fsub fast float [[TMP1]], [[TMP2]] 14 %1 = fmul fast float %x, %y 15 %2 = fmul fast float %y, %x 16 %3 = fsub fast float %1, %2 24 ; CHECK-NEXT: [[TMP1:%.*]] = fmul fast float %y, %x 25 ; CHECK-NEXT: [[TMP2:%.*]] = fmul fast float %y, %x 26 ; CHECK-NEXT: [[TMP3:%.*]] = fsub fast float [[TMP1]], [[TMP2]] 29 %1 = fmul fast float %x, %y [all …]
|
D | reassoc-intermediate-fnegs.ll | 8 ; CHECK-NEXT: [[T2_NEG:%.*]] = fmul fast half [[A:%.*]], 0xH4500 9 ; CHECK-NEXT: [[REASS_MUL:%.*]] = fmul fast half [[B:%.*]], 0xH4500 10 ; CHECK-NEXT: [[T51:%.*]] = fsub fast half [[REASS_MUL]], [[T2_NEG]] 11 ; CHECK-NEXT: [[T5:%.*]] = fadd fast half [[REASS_MUL]], [[T2_NEG]] 14 %t1 = fmul fast half %b, 0xH4200 ; 3*b 15 %t2 = fmul fast half %a, 0xH4500 ; 5*a 16 %t3 = fmul fast half %b, 0xH4000 ; 2*b 17 %t4 = fsub fast half %t2, %t1 ; 5 * a - 3 * b 18 %t5 = fsub fast half %t3, %t4 ; 2 * b - ( 5 * a - 3 * b) 26 ; CHECK-NEXT: [[T2:%.*]] = fmul fast half [[A:%.*]], 0xH4500 [all …]
|
D | fast-multistep.ll | 7 ; CHECK-NEXT: [[REASS_ADD1:%.*]] = fadd fast float %c, %b 8 ; CHECK-NEXT: [[REASS_MUL2:%.*]] = fmul fast float %a, %a 9 ; CHECK-NEXT: [[REASS_MUL:%.*]] = fmul fast float [[REASS_MUL:%.*]]2, [[REASS_ADD1]] 12 %t0 = fmul fast float %a, %b 13 %t1 = fmul fast float %a, %t0 ; a*(a*b) 14 %t2 = fmul fast float %a, %c 15 %t3 = fmul fast float %a, %t2 ; a*(a*c) 16 %t4 = fadd fast float %t1, %t3 24 ; CHECK-NEXT: [[REASS_ADD:%.*]] = fadd fast float %c, %b 25 ; CHECK-NEXT: [[REASS_MUL:%.*]] = fmul fast float [[REASS_ADD]], %a [all …]
|
D | mixed-fast-nonfast-fp.ll | 6 ; CHECK-NEXT: [[FACTOR:%.*]] = fmul fast float %c, 2.000000e+00 7 ; CHECK-NEXT: [[REASS_ADD1:%.*]] = fadd fast float [[FACTOR]], %b 8 ; CHECK-NEXT: [[REASS_MUL:%.*]] = fmul fast float [[REASS_ADD1]], %a 9 ; CHECK-NEXT: [[ADD3:%.*]] = fadd fast float [[REASS_MUL]], [[MUL3]] 12 %mul1 = fmul fast float %a, %c 13 %mul2 = fmul fast float %a, %b 15 %mul4 = fmul fast float %a, %c 16 %add1 = fadd fast float %mul1, %mul3 17 %add2 = fadd fast float %mul4, %mul2 18 %add3 = fadd fast float %add1, %add2 [all …]
|
/external/llvm/test/Transforms/InstCombine/ |
D | fast-math.ll | 6 %mul = fmul fast float %a, 0x3FF3333340000000 7 %mul1 = fmul fast float %mul, 0x4002666660000000 10 ; CHECK: fmul fast float %a, 0x4006147AE0000000 17 ; CHECK: %mul = fmul fast float %a, 0x3FF3333340000000 18 %mul = fmul fast float %a, 0x3FF3333340000000 25 ; CHECK: fmul fast float %a, 0x4006147AE0000000 27 %mul1 = fmul fast float %mul, 0x4002666660000000 33 %t1 = fmul fast double 2.000000e+00, %f1 34 %t2 = fadd fast double %f1, %t1 37 ; CHECK: fmul fast double %f1, 3.000000e+00 [all …]
|
D | double-float-shrink-1.ll | 14 %call = call fast double @acos(double %conv) 18 ; CHECK: call fast float @acosf(float %f) 23 %call = call fast double @acos(double %conv) 26 ; CHECK: call fast double @acos(double %conv) 31 %call = call fast double @acosh(double %conv) 35 ; CHECK: call fast float @acoshf(float %f) 40 %call = call fast double @acosh(double %conv) 43 ; CHECK: call fast double @acosh(double %conv) 48 %call = call fast double @asin(double %conv) 52 ; CHECK: call fast float @asinf(float %f) [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/InstCombine/ |
D | pow-4.ll | 10 ; CHECK-NEXT: [[TMP1:%.*]] = fmul fast float [[X:%.*]], [[X]] 11 ; CHECK-NEXT: [[TMP2:%.*]] = fmul fast float [[TMP1]], [[TMP1]] 14 %1 = call fast float @llvm.pow.f32(float %x, float 4.000000e+00) 21 ; CHECK-NEXT: [[TMP1:%.*]] = fmul fast double [[X:%.*]], [[X]] 22 ; CHECK-NEXT: [[TMP2:%.*]] = fmul fast double [[TMP1]], [[X]] 25 %1 = call fast double @llvm.pow.f64(double %x, double 3.000000e+00) 32 ; CHECK-NEXT: [[TMP1:%.*]] = fmul fast double [[X:%.*]], [[X]] 33 ; CHECK-NEXT: [[TMP2:%.*]] = fmul fast double [[TMP1]], [[TMP1]] 36 %1 = call fast double @llvm.pow.f64(double %x, double 4.000000e+00) 43 ; CHECK-NEXT: [[TMP1:%.*]] = fmul fast double [[X:%.*]], [[X]] [all …]
|
D | fast-math.ll | 8 ; CHECK-NEXT: [[MUL1:%.*]] = fmul fast float [[A:%.*]], 0x4006147AE0000000 11 %mul = fmul fast float %a, 0x3FF3333340000000 12 %mul1 = fmul fast float %mul, 0x4002666660000000 20 ; CHECK-NEXT: [[MUL:%.*]] = fmul fast float [[A:%.*]], 0x3FF3333340000000 24 %mul = fmul fast float %a, 0x3FF3333340000000 31 ; CHECK-NEXT: [[MUL1:%.*]] = fmul fast float [[A:%.*]], 0x4006147AE0000000 35 %mul1 = fmul fast float %mul, 0x4002666660000000 45 ; CHECK-NEXT: [[TMP1:%.*]] = fmul fast double [[F1:%.*]], 6.000000e+00 48 %t1 = fmul fast double 5.000000e+00, %f1 49 %t2 = fadd fast double %f1, %t1 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/SLPVectorizer/X86/ |
D | horizontal-list.ll | 18 ; CHECK-NEXT: [[TMP3:%.*]] = fmul fast <2 x float> [[TMP2]], [[TMP1]] 20 ; CHECK-NEXT: [[ADD:%.*]] = fadd fast float [[TMP4]], [[CONV]] 22 ; CHECK-NEXT: [[ADD_1:%.*]] = fadd fast float [[TMP5]], [[ADD]] 25 ; CHECK-NEXT: [[TMP8:%.*]] = fmul fast <2 x float> [[TMP7]], [[TMP6]] 27 ; CHECK-NEXT: [[ADD_2:%.*]] = fadd fast float [[TMP9]], [[ADD_1]] 29 ; CHECK-NEXT: [[ADD_3:%.*]] = fadd fast float [[TMP10]], [[ADD_2]] 30 ; CHECK-NEXT: [[ADD7:%.*]] = fadd fast float [[ADD_3]], [[CONV]] 31 ; CHECK-NEXT: [[ADD19:%.*]] = fadd fast float [[TMP4]], [[ADD7]] 32 ; CHECK-NEXT: [[ADD19_1:%.*]] = fadd fast float [[TMP5]], [[ADD19]] 33 ; CHECK-NEXT: [[ADD19_2:%.*]] = fadd fast float [[TMP9]], [[ADD19_1]] [all …]
|
D | reverse_extract_elements.ll | 7 ; CHECK-NEXT: [[TMP0:%.*]] = fmul fast <4 x float> [[X:%.*]], [[Y:%.*]] 9 ; CHECK-NEXT: [[BIN_RDX:%.*]] = fadd fast <4 x float> [[TMP0]], [[RDX_SHUF]] 11 ; CHECK-NEXT: [[BIN_RDX2:%.*]] = fadd fast <4 x float> [[BIN_RDX]], [[RDX_SHUF1]] 18 %mul = fmul fast float %vecext, %vecext1 21 %mul.1 = fmul fast float %vecext.1, %vecext1.1 22 %add.1 = fadd fast float %mul.1, %mul 25 %mul.2 = fmul fast float %vecext.2, %vecext1.2 26 %add.2 = fadd fast float %mul.2, %add.1 29 %mul.3 = fmul fast float %vecext.3, %vecext1.3 30 %add.3 = fadd fast float %mul.3, %add.2 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AArch64/ |
D | PBQP-chain.ll | 1 ; RUN: llc < %s -mcpu=cortex-a57 -mattr=+neon -fp-contract=fast -regalloc=pbqp -pbqp-coalescing | F… 2 ; RUN: llc < %s -mcpu=cortex-a57 -mattr=+neon -fp-contract=fast -regalloc=pbqp -pbqp-coalescing | F… 27 %mul = fmul fast double %1, %0 29 %mul7 = fmul fast double %2, %0 34 %mul.1 = fmul fast double %4, %3 35 %add.1 = fadd fast double %mul.1, %mul 38 %mul7.1 = fmul fast double %5, %3 39 %add8.1 = fadd fast double %mul7.1, %mul7 44 %mul.2 = fmul fast double %7, %6 45 %add.2 = fadd fast double %mul.2, %add.1 [all …]
|
D | aarch64-a57-fp-load-balancing.ll | 47 %mul = fmul fast double %0, %1 48 %add = fadd fast double %mul, %4 49 %mul5 = fmul fast double %1, %2 50 %add6 = fadd fast double %mul5, %add 51 %mul7 = fmul fast double %1, %3 52 %sub = fsub fast double %add6, %mul7 53 %mul8 = fmul fast double %2, %3 54 %add9 = fadd fast double %mul8, %sub 62 %mul15 = fmul fast double %6, %7 63 %mul16 = fmul fast double %0, %5 [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | PBQP-chain.ll | 1 ; RUN: llc < %s -mcpu=cortex-a57 -mattr=+neon -fp-contract=fast -regalloc=pbqp -pbqp-coalescing | F… 2 ; RUN: llc < %s -mcpu=cortex-a57 -mattr=+neon -fp-contract=fast -regalloc=pbqp -pbqp-coalescing | F… 27 %mul = fmul fast double %1, %0 29 %mul7 = fmul fast double %2, %0 34 %mul.1 = fmul fast double %4, %3 35 %add.1 = fadd fast double %mul.1, %mul 38 %mul7.1 = fmul fast double %5, %3 39 %add8.1 = fadd fast double %mul7.1, %mul7 44 %mul.2 = fmul fast double %7, %6 45 %add.2 = fadd fast double %mul.2, %add.1 [all …]
|
D | aarch64-a57-fp-load-balancing.ll | 47 %mul = fmul fast double %0, %1 48 %add = fadd fast double %mul, %4 49 %mul5 = fmul fast double %1, %2 50 %add6 = fadd fast double %mul5, %add 51 %mul7 = fmul fast double %1, %3 52 %sub = fsub fast double %add6, %mul7 53 %mul8 = fmul fast double %2, %3 54 %add9 = fadd fast double %mul8, %sub 62 %mul15 = fmul fast double %6, %7 63 %mul16 = fmul fast double %0, %5 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AMDGPU/ |
D | simplify-libcalls.ll | 6 ; GCN-POSTLINK: tail call fast float @_Z3sinf( 7 ; GCN-POSTLINK: tail call fast float @_Z3cosf( 8 ; GCN-PRELINK: call fast float @_Z6sincosfPf( 9 ; GCN-NATIVE: tail call fast float @_Z10native_sinf( 10 ; GCN-NATIVE: tail call fast float @_Z10native_cosf( 14 %call = tail call fast float @_Z3sinf(float %tmp) 16 %call2 = tail call fast float @_Z3cosf(float %tmp) 27 ; GCN-POSTLINK: tail call fast <2 x float> @_Z3sinDv2_f( 28 ; GCN-POSTLINK: tail call fast <2 x float> @_Z3cosDv2_f( 29 ; GCN-PRELINK: call fast <2 x float> @_Z6sincosDv2_fPS_( [all …]
|
D | amdgpu-codegenprepare-fdiv.ll | 17 ; CHECK: %md.25ulp = call float @llvm.amdgcn.fdiv.fast(float %a, float %b), !fpmath !0 18 ; CHECK: %md.3ulp = call float @llvm.amdgcn.fdiv.fast(float %a, float %b), !fpmath !3 19 ; CHECK: %fast.md.25ulp = fdiv fast float %a, %b, !fpmath !0 37 %fast.md.25ulp = fdiv fast float %a, %b, !fpmath !0 38 store volatile float %fast.md.25ulp, float addrspace(1)* %out 52 ; CHECK: %fast.no.md = fdiv fast float 1.000000e+00, %x{{$}} 53 ; CHECK: %fast.25ulp = fdiv fast float 1.000000e+00, %x, !fpmath !0 70 %fast.no.md = fdiv fast float 1.0, %x 71 store volatile float %fast.no.md, float addrspace(1)* %out 73 %fast.25ulp = fdiv fast float 1.0, %x, !fpmath !0 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/Transforms/SimplifyCFG/AArch64/ |
D | prefer-fma.ll | 16 ; CHECK: %3 = fmul fast double %1, %2 17 ; CHECK-NEXT: %mul = fadd fast double 1.000000e+00, %3 19 %3 = fmul fast double %1, %2 20 %mul = fadd fast double 1.000000e+00, %3 27 ; CHECK: %mul1 = fmul fast double %1, %2 28 ; CHECK-NEXT: %sub1 = fsub fast double %mul1, %0 30 %mul1 = fmul fast double %1, %4 31 %sub1 = fsub fast double %mul1, %0 45 ; CHECK: %7 = fmul fast double %5, 3.000000e+00 46 ; CHECK-NEXT: %mul2 = fsub fast double %6, %7 [all …]
|
/external/llvm/test/Transforms/SimplifyCFG/AArch64/ |
D | prefer-fma.ll | 16 ; CHECK: %3 = fmul fast double %1, %2 17 ; CHECK-NEXT: %mul = fadd fast double 1.000000e+00, %3 19 %3 = fmul fast double %1, %2 20 %mul = fadd fast double 1.000000e+00, %3 27 ; CHECK: %mul1 = fmul fast double %1, %2 28 ; CHECK-NEXT: %sub1 = fsub fast double %mul1, %0 30 %mul1 = fmul fast double %1, %4 31 %sub1 = fsub fast double %mul1, %0 44 ; CHECK: %7 = fmul fast double %5, 3.000000e+00 45 ; CHECK-NEXT: %mul2 = fsub fast double %6, %7 [all …]
|