/external/llvm/test/Transforms/StraightLineStrengthReduce/ |
D | slsr-mul.ll | 15 %mul1 = mul i32 %b1, %s 16 call void @foo(i32 %mul1) 36 %mul1 = mul i32 %b1, %s 37 call void @foo(i32 %mul1) 57 %mul1 = mul i32 %b1, %s 59 call void @foo(i32 %mul1) 81 %mul1 = mul i32 %a1, %b 86 call void @foo(i32 %mul1) 101 ; mul1 = mul0 + bump; // GVN ensures mul1 and mul2 use the same bump. 102 ; mul2 = mul1 + bump; [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | shifter_operand.ll | 85 %mul1 = mul i32 %y, 127534 86 %or = or i32 %mul1, %x 97 %mul1 = mul i32 %y, 255068 98 %or = or i32 %mul1, %x 109 %mul1 = mul i32 %y, 510136 110 %or = or i32 %mul1, %x 121 %mul1 = mul i32 %y, 1020272 122 %or = or i32 %mul1, %x 159 %mul1 = mul i32 %y, 127534 160 %arrayidx = getelementptr inbounds i8, i8* %x, i32 %mul1 [all …]
|
D | 2009-04-08-FloatUndef.ll | 7 %mul1 = fmul <4 x float> %shuffle7, zeroinitializer ; <<4 x float>> [#uses=1] 8 %add2 = fadd <4 x float> %mul1, %input2 ; <<4 x float>> [#uses=1]
|
/external/boringssl/src/crypto/bn/asm/ |
D | x86-mont.pl | 112 $mul1="mm5"; 131 &movd ($mul1,&DWP(0,$ap)); # ap[0] 134 &pmuludq($mul1,$mul0); # ap[0]*bp[0] 135 &movq ($car0,$mul1); 136 &movq ($acc0,$mul1); # I wish movd worked for 139 &pmuludq($mul1,$_n0q); # *=n0 141 &pmuludq($car1,$mul1); # "t[0]"*np[0]*n0 153 &pmuludq($acc1,$mul1); # np[j]*m1 171 &pmuludq($acc1,$mul1); # np[num-1]*m1 191 &movd ($mul1,&DWP(0,$ap)); # ap[0] [all …]
|
/external/llvm/test/Transforms/GVN/ |
D | 2011-07-07-MatchIntrinsicExtract.ll | 34 %mul1 = mul i64 %a, %b 35 ret i64 %mul1 39 ; CHECK-NOT: mul1 70 %mul1 = mul i64 %a, %b 71 ret i64 %mul1 75 ; CHECK-NOT: mul1
|
/external/llvm/test/Transforms/InstCombine/ |
D | add2.ll | 218 %mul1 = mul nsw i16 %x, 8 219 %add2 = add nsw i16 %x, %mul1 227 %mul1 = mul nsw i16 %x, 8 228 %add2 = add nsw i16 %mul1, %x 236 %mul1 = mul i16 %a, 2 238 %add = add nsw i16 %mul1, %mul2 246 %mul1 = mul nsw i16 %a, 2 248 %add = add nsw i16 %mul1, %mul2 256 %mul1 = mul nsw i16 %a, 3 258 %add = add nsw i16 %mul1, %mul2 [all …]
|
/external/vulkan-validation-layers/libs/glm/detail/ |
D | intrinsic_geometric.inl | 82 __m128 mul1 = _mm_mul_ps(swp1, swp2); local 83 __m128 sub0 = _mm_sub_ps(mul0, mul1); 108 __m128 mul1 = _mm_mul_ps(N, mul0); local 109 return mul1; 117 __m128 mul1 = _mm_mul_ps(mul0, glm::detail::two); local 118 __m128 sub0 = _mm_sub_ps(I, mul1); 127 __m128 mul1 = _mm_mul_ps(dot0, dot0); local 129 __m128 sub1 = _mm_sub_ps(glm::detail::one, mul1);
|
/external/llvm/test/CodeGen/X86/ |
D | dag-fmf-cse.ll | 4 ; If fast-math-flags are propagated correctly, the mul1 expression 16 %mul1 = fmul fast float %a, %b 19 %abx2 = fsub fast float %mul1, %mul2
|
D | atom-sched.ll | 30 %mul1 = mul nsw i32 %2, %3 31 store i32 %mul1, i32* @d, align 4
|
D | i128-mul.ll | 16 ; CHECK: @mul1 20 define i64 @mul1(i64 %n, i64* nocapture %z, i64* nocapture %x, i64 %y) {
|
/external/llvm/test/Transforms/SimplifyCFG/AArch64/ |
D | prefer-fma.ll | 27 ; CHECK: %mul1 = fmul fast double %1, %2 28 ; CHECK-NEXT: %sub1 = fsub fast double %mul1, %0 30 %mul1 = fmul fast double %1, %4 31 %sub1 = fsub fast double %mul1, %0
|
/external/llvm/test/Transforms/Reassociate/ |
D | canonicalize-neg-const.ll | 14 %mul1 = fmul double -1.234000e-01, %y 15 %add2 = fadd double %mul1, %x 30 %mul1 = fmul double %y, -1.234000e-01 31 %add2 = fadd double %mul1, %x 46 %mul1 = fmul double %y, -1.234000e-01 47 %add2 = fsub double %x, %mul1
|
D | mixed-fast-nonfast-fp.ll | 10 %mul1 = fmul fast float %a, %c 14 %add1 = fadd fast float %mul1, %mul3
|
/external/llvm/test/Analysis/LoopAccessAnalysis/ |
D | safe-no-checks.ll | 17 ; CHECK-NEXT: store i16 %mul1, i16* %arrayidxA, align 2 41 %mul1 = mul i16 %mul, %loadC 44 store i16 %mul1, i16* %arrayidxA, align 2
|
D | unsafe-and-rt-checks.ll | 14 ; CHECK-NEXT: store i16 %mul1, i16* %arrayidxA_plus_2, align 2 53 %mul1 = mul i16 %mul, %loadC 57 store i16 %mul1, i16* %arrayidxA_plus_2, align 2
|
D | number-of-memchecks.ll | 43 %mul1 = mul i16 %mul, %loadC 46 store i16 %mul1, i16* %arrayidxD, align 2 52 store i16 %mul1, i16* %arrayidxF, align 2 133 %mul1 = mul i16 %mul, %loadB 136 store i16 %mul1, i16* %arrayidxC, align 2 205 %mul1 = mul i16 %mul, %loadB 208 store i16 %mul1, i16* %arrayidxC, align 2
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-fp-contract-zero.ll | 11 %mul1 = fmul double %b, 0.000000e+00 12 %sub = fsub double %mul, %mul1
|
/external/llvm/test/CodeGen/PowerPC/ |
D | 2007-11-19-VectorSplitting.ll | 13 %mul1 = fmul <4 x float> %shuffle7, zeroinitializer ; <<4 x 14 %add2 = fadd <4 x float> %mul1, %input2 ; <<4 x float>>
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | idct_msa.c | 182 v8i16 input0, input1, dequant_in0, dequant_in1, mul0, mul1; in dequant_idct4x4_addblk_msa() local 194 MUL2(input0, dequant_in0, input1, dequant_in1, mul0, mul1); in dequant_idct4x4_addblk_msa() 195 PCKEV_D2_SH(zero, mul0, zero, mul1, in0, in2); in dequant_idct4x4_addblk_msa() 196 PCKOD_D2_SH(zero, mul0, zero, mul1, in1, in3); in dequant_idct4x4_addblk_msa() 198 PCKEV_D2_SH(hz1_h, hz0_h, hz3_h, hz2_h, mul0, mul1); in dequant_idct4x4_addblk_msa() 200 UNPCK_SH_SW(mul1, hz2_w, hz3_w); in dequant_idct4x4_addblk_msa() 226 v8i16 mul0, mul1, mul2, mul3, dequant_in0, dequant_in1; in dequant_idct4x4_addblk_2x_msa() local 236 mul0, mul1, mul2, mul3); in dequant_idct4x4_addblk_2x_msa() 237 PCKEV_D2_SH(mul2, mul0, mul3, mul1, in0, in2); in dequant_idct4x4_addblk_2x_msa() 238 PCKOD_D2_SH(mul2, mul0, mul3, mul1, in1, in3); in dequant_idct4x4_addblk_2x_msa()
|
/external/llvm/test/Transforms/Inline/ |
D | vector-bonus.ll | 9 %mul1 = mul nsw i32 %i, %i 25 %retval.0 = phi i32 [ %mul1, %if.then ], [ %add7, %if.else ]
|
/external/llvm/test/Transforms/SLPVectorizer/X86/ |
D | barriercall.ll | 15 %mul1 = mul nsw i32 %n, 9 16 %add2 = add nsw i32 %mul1, 9
|
D | extractcost.ll | 14 %mul1 = mul nsw i32 %n, 9 15 %add2 = add nsw i32 %mul1, 9
|
D | crash_binaryop.ll | 21 %mul1 = fmul double %add1, 0.000000e+00 30 %binary_V = fadd double %mul1, %binaryop_B
|
/external/llvm/test/Analysis/DependenceAnalysis/ |
D | SymbolicSIV.ll | 77 %mul1 = mul i64 %n, 5 78 %add = add i64 %mul, %mul1 215 %mul1 = mul i64 %n, 3 216 %add = add i64 %mul, %mul1 260 %mul1 = shl i64 %n, 1 261 %sub = sub i64 %mul, %mul1 352 %mul1 = mul i64 %mul, %i.03 353 %add = add i64 %mul1, %M 401 %mul1 = mul i64 %mul, %i.03 402 %add = add i64 %mul1, %M
|
/external/llvm/test/Transforms/LoopIdiom/AMDGPU/ |
D | popcnt.ll | 85 %mydata2.addr.012 = phi i32 [ %mul1, %while.body ], [ %mydata2, %entry ] 93 %mul1 = mul nsw i32 %conv, %mydata2.addr.012 99 %mydata2.addr.0.lcssa = phi i32 [ %mydata2, %entry ], [ %mul1, %while.body ]
|