/external/pdfium/core/src/fxcodec/jbig2/ |
D | JBig2_Image.cpp | 244 tmp1 = 0, tmp2 = 0, maskL = 0, maskR = 0, maskM = 0; in composeTo_opt2() local 298 tmp2 = JBIG2_GETDWORD(lineDst); in composeTo_opt2() 301 tmp = (tmp2 & ~maskM) | ((tmp1 | tmp2) & maskM); in composeTo_opt2() 304 tmp = (tmp2 & ~maskM) | ((tmp1 & tmp2) & maskM); in composeTo_opt2() 307 tmp = (tmp2 & ~maskM) | ((tmp1 ^ tmp2) & maskM); in composeTo_opt2() 310 tmp = (tmp2 & ~maskM) | ((~(tmp1 ^ tmp2)) & maskM); in composeTo_opt2() 313 tmp = (tmp2 & ~maskM) | (tmp1 & maskM); in composeTo_opt2() 327 tmp2 = JBIG2_GETDWORD(lineDst); in composeTo_opt2() 330 tmp = (tmp2 & ~maskM) | ((tmp1 | tmp2) & maskM); in composeTo_opt2() 333 tmp = (tmp2 & ~maskM) | ((tmp1 & tmp2) & maskM); in composeTo_opt2() [all …]
|
/external/webrtc/webrtc/modules/audio_coding/codecs/isac/fix/source/ |
D | lpc_masking_model_mips.c | 36 int32_t tmp2, tmp3; in WebRtcIsacfix_CalculateResidualEnergyMIPS() local 55 : [tmp2] "=&r" (tmp2), [tmp3] "=&r" (tmp3), [tmp32] "=&r" (tmp32), in WebRtcIsacfix_CalculateResidualEnergyMIPS() 71 : [tmp2] "r" (tmp2), [tmp3] "r" (tmp3) in WebRtcIsacfix_CalculateResidualEnergyMIPS() 75 if (((!(sign_1 || sign_2)) && (0x7FFFFFFF - sum64_hi < tmp2)) || in WebRtcIsacfix_CalculateResidualEnergyMIPS() 76 ((sign_1 && sign_2) && (sum64_hi + tmp2 > 0))) { in WebRtcIsacfix_CalculateResidualEnergyMIPS() 89 : [tmp2] "+r" (tmp2), [tmp3] "+r" (tmp3), in WebRtcIsacfix_CalculateResidualEnergyMIPS() 103 : [tmp2] "r" (tmp2), [tmp3] "r" (tmp3) in WebRtcIsacfix_CalculateResidualEnergyMIPS() 119 int32_t tmp2, tmp3; in WebRtcIsacfix_CalculateResidualEnergyMIPS() local 141 : [tmp2] "=&r" (tmp2), [tmp3] "=&r" (tmp3), [tmp32] "=&r" (tmp32), in WebRtcIsacfix_CalculateResidualEnergyMIPS() 156 : [tmp2] "+r" (tmp2), [tmp3] "+r" (tmp3), [sum64_hi] "+r" (sum64_hi), in WebRtcIsacfix_CalculateResidualEnergyMIPS() [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | vshift.ll | 7 %tmp2 = load <8 x i8>, <8 x i8>* %B 8 %tmp3 = shl <8 x i8> %tmp1, %tmp2 16 %tmp2 = load <4 x i16>, <4 x i16>* %B 17 %tmp3 = shl <4 x i16> %tmp1, %tmp2 25 %tmp2 = load <2 x i32>, <2 x i32>* %B 26 %tmp3 = shl <2 x i32> %tmp1, %tmp2 34 %tmp2 = load <1 x i64>, <1 x i64>* %B 35 %tmp3 = shl <1 x i64> %tmp1, %tmp2 43 %tmp2 = shl <8 x i8> %tmp1, < i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7 > 44 ret <8 x i8> %tmp2 [all …]
|
D | vbits.ll | 7 %tmp2 = load <8 x i8>, <8 x i8>* %B 8 %tmp3 = and <8 x i8> %tmp1, %tmp2 16 %tmp2 = load <4 x i16>, <4 x i16>* %B 17 %tmp3 = and <4 x i16> %tmp1, %tmp2 25 %tmp2 = load <2 x i32>, <2 x i32>* %B 26 %tmp3 = and <2 x i32> %tmp1, %tmp2 34 %tmp2 = load <1 x i64>, <1 x i64>* %B 35 %tmp3 = and <1 x i64> %tmp1, %tmp2 43 %tmp2 = load <16 x i8>, <16 x i8>* %B 44 %tmp3 = and <16 x i8> %tmp1, %tmp2 [all …]
|
D | vneg.ll | 7 %tmp2 = sub <8 x i8> zeroinitializer, %tmp1 8 ret <8 x i8> %tmp2 15 %tmp2 = sub <4 x i16> zeroinitializer, %tmp1 16 ret <4 x i16> %tmp2 23 %tmp2 = sub <2 x i32> zeroinitializer, %tmp1 24 ret <2 x i32> %tmp2 31 %tmp2 = fsub <2 x float> < float -0.000000e+00, float -0.000000e+00 >, %tmp1 32 ret <2 x float> %tmp2 39 %tmp2 = sub <16 x i8> zeroinitializer, %tmp1 40 ret <16 x i8> %tmp2 [all …]
|
D | vshl.ll | 7 %tmp2 = load <8 x i8>, <8 x i8>* %B 8 %tmp3 = call <8 x i8> @llvm.arm.neon.vshifts.v8i8(<8 x i8> %tmp1, <8 x i8> %tmp2) 16 %tmp2 = load <4 x i16>, <4 x i16>* %B 17 %tmp3 = call <4 x i16> @llvm.arm.neon.vshifts.v4i16(<4 x i16> %tmp1, <4 x i16> %tmp2) 25 %tmp2 = load <2 x i32>, <2 x i32>* %B 26 %tmp3 = call <2 x i32> @llvm.arm.neon.vshifts.v2i32(<2 x i32> %tmp1, <2 x i32> %tmp2) 34 %tmp2 = load <1 x i64>, <1 x i64>* %B 35 %tmp3 = call <1 x i64> @llvm.arm.neon.vshifts.v1i64(<1 x i64> %tmp1, <1 x i64> %tmp2) 43 %tmp2 = load <8 x i8>, <8 x i8>* %B 44 %tmp3 = call <8 x i8> @llvm.arm.neon.vshiftu.v8i8(<8 x i8> %tmp1, <8 x i8> %tmp2) [all …]
|
D | vqshl.ll | 7 %tmp2 = load <8 x i8>, <8 x i8>* %B 8 %tmp3 = call <8 x i8> @llvm.arm.neon.vqshifts.v8i8(<8 x i8> %tmp1, <8 x i8> %tmp2) 16 %tmp2 = load <4 x i16>, <4 x i16>* %B 17 %tmp3 = call <4 x i16> @llvm.arm.neon.vqshifts.v4i16(<4 x i16> %tmp1, <4 x i16> %tmp2) 25 %tmp2 = load <2 x i32>, <2 x i32>* %B 26 %tmp3 = call <2 x i32> @llvm.arm.neon.vqshifts.v2i32(<2 x i32> %tmp1, <2 x i32> %tmp2) 34 %tmp2 = load <1 x i64>, <1 x i64>* %B 35 %tmp3 = call <1 x i64> @llvm.arm.neon.vqshifts.v1i64(<1 x i64> %tmp1, <1 x i64> %tmp2) 43 %tmp2 = load <8 x i8>, <8 x i8>* %B 44 %tmp3 = call <8 x i8> @llvm.arm.neon.vqshiftu.v8i8(<8 x i8> %tmp1, <8 x i8> %tmp2) [all …]
|
D | vrev.ll | 7 …%tmp2 = shufflevector <8 x i8> %tmp1, <8 x i8> undef, <8 x i32> <i32 7, i32 6, i32 5, i32 4, i32 3… 8 ret <8 x i8> %tmp2 15 %tmp2 = shufflevector <4 x i16> %tmp1, <4 x i16> undef, <4 x i32> <i32 3, i32 2, i32 1, i32 0> 16 ret <4 x i16> %tmp2 23 %tmp2 = shufflevector <2 x i32> %tmp1, <2 x i32> undef, <2 x i32> <i32 1, i32 0> 24 ret <2 x i32> %tmp2 31 %tmp2 = shufflevector <2 x float> %tmp1, <2 x float> undef, <2 x i32> <i32 1, i32 0> 32 ret <2 x float> %tmp2 39 …%tmp2 = shufflevector <16 x i8> %tmp1, <16 x i8> undef, <16 x i32> <i32 7, i32 6, i32 5, i32 4, i3… 40 ret <16 x i8> %tmp2 [all …]
|
D | vadd.ll | 7 %tmp2 = load <8 x i8>, <8 x i8>* %B 8 %tmp3 = add <8 x i8> %tmp1, %tmp2 16 %tmp2 = load <4 x i16>, <4 x i16>* %B 17 %tmp3 = add <4 x i16> %tmp1, %tmp2 25 %tmp2 = load <2 x i32>, <2 x i32>* %B 26 %tmp3 = add <2 x i32> %tmp1, %tmp2 34 %tmp2 = load <1 x i64>, <1 x i64>* %B 35 %tmp3 = add <1 x i64> %tmp1, %tmp2 43 %tmp2 = load <2 x float>, <2 x float>* %B 44 %tmp3 = fadd <2 x float> %tmp1, %tmp2 [all …]
|
D | vsub.ll | 7 %tmp2 = load <8 x i8>, <8 x i8>* %B 8 %tmp3 = sub <8 x i8> %tmp1, %tmp2 16 %tmp2 = load <4 x i16>, <4 x i16>* %B 17 %tmp3 = sub <4 x i16> %tmp1, %tmp2 25 %tmp2 = load <2 x i32>, <2 x i32>* %B 26 %tmp3 = sub <2 x i32> %tmp1, %tmp2 34 %tmp2 = load <1 x i64>, <1 x i64>* %B 35 %tmp3 = sub <1 x i64> %tmp1, %tmp2 43 %tmp2 = load <2 x float>, <2 x float>* %B 44 %tmp3 = fsub <2 x float> %tmp1, %tmp2 [all …]
|
D | vldlane.ll | 11 %tmp2 = load i8, i8* %A, align 8 12 %tmp3 = insertelement <8 x i8> %tmp1, i8 %tmp2, i32 3 21 %tmp2 = load i16, i16* %A, align 8 22 %tmp3 = insertelement <4 x i16> %tmp1, i16 %tmp2, i32 2 31 %tmp2 = load i32, i32* %A, align 8 32 %tmp3 = insertelement <2 x i32> %tmp1, i32 %tmp2, i32 1 41 %tmp2 = load i32, i32* %A, align 4 42 %tmp3 = insertelement <2 x i32> %tmp1, i32 %tmp2, i32 1 50 %tmp2 = load float, float* %A, align 4 51 %tmp3 = insertelement <2 x float> %tmp1, float %tmp2, i32 1 [all …]
|
D | vget_lane.ll | 9 %tmp2 = extractelement <8 x i8> %tmp1, i32 1 10 %tmp3 = sext i8 %tmp2 to i32 18 %tmp2 = extractelement <4 x i16> %tmp1, i32 1 19 %tmp3 = sext i16 %tmp2 to i32 27 %tmp2 = extractelement <8 x i8> %tmp1, i32 1 28 %tmp3 = zext i8 %tmp2 to i32 36 %tmp2 = extractelement <4 x i16> %tmp1, i32 1 37 %tmp3 = zext i16 %tmp2 to i32 46 %tmp2 = add <2 x i32> %tmp1, %tmp1 47 %tmp3 = extractelement <2 x i32> %tmp2, i32 1 [all …]
|
D | vcvt.ll | 7 %tmp2 = fptosi <2 x float> %tmp1 to <2 x i32> 8 ret <2 x i32> %tmp2 15 %tmp2 = fptoui <2 x float> %tmp1 to <2 x i32> 16 ret <2 x i32> %tmp2 23 %tmp2 = sitofp <2 x i32> %tmp1 to <2 x float> 24 ret <2 x float> %tmp2 31 %tmp2 = uitofp <2 x i32> %tmp1 to <2 x float> 32 ret <2 x float> %tmp2 39 %tmp2 = fptosi <4 x float> %tmp1 to <4 x i32> 40 ret <4 x i32> %tmp2 [all …]
|
D | popcnt.ll | 8 %tmp2 = call <8 x i8> @llvm.ctpop.v8i8(<8 x i8> %tmp1) 9 ret <8 x i8> %tmp2 16 %tmp2 = call <16 x i8> @llvm.ctpop.v16i8(<16 x i8> %tmp1) 17 ret <16 x i8> %tmp2 28 %tmp2 = call <4 x i16> @llvm.ctpop.v4i16(<4 x i16> %tmp1) 29 ret <4 x i16> %tmp2 40 %tmp2 = call <8 x i16> @llvm.ctpop.v8i16(<8 x i16> %tmp1) 41 ret <8 x i16> %tmp2 55 %tmp2 = call <2 x i32> @llvm.ctpop.v2i32(<2 x i32> %tmp1) 56 ret <2 x i32> %tmp2 [all …]
|
D | vcvt-v8.ll | 6 %tmp2 = call <4 x i32> @llvm.arm.neon.vcvtas.v4i32.v4f32(<4 x float> %tmp1) 7 ret <4 x i32> %tmp2 14 %tmp2 = call <2 x i32> @llvm.arm.neon.vcvtas.v2i32.v2f32(<2 x float> %tmp1) 15 ret <2 x i32> %tmp2 22 %tmp2 = call <4 x i32> @llvm.arm.neon.vcvtns.v4i32.v4f32(<4 x float> %tmp1) 23 ret <4 x i32> %tmp2 30 %tmp2 = call <2 x i32> @llvm.arm.neon.vcvtns.v2i32.v2f32(<2 x float> %tmp1) 31 ret <2 x i32> %tmp2 38 %tmp2 = call <4 x i32> @llvm.arm.neon.vcvtps.v4i32.v4f32(<4 x float> %tmp1) 39 ret <4 x i32> %tmp2 [all …]
|
D | vcnt.ll | 8 %tmp2 = call <8 x i8> @llvm.ctpop.v8i8(<8 x i8> %tmp1) 9 ret <8 x i8> %tmp2 16 %tmp2 = call <16 x i8> @llvm.ctpop.v16i8(<16 x i8> %tmp1) 17 ret <16 x i8> %tmp2 27 %tmp2 = call <8 x i8> @llvm.ctlz.v8i8(<8 x i8> %tmp1, i1 0) 28 ret <8 x i8> %tmp2 35 %tmp2 = call <4 x i16> @llvm.ctlz.v4i16(<4 x i16> %tmp1, i1 0) 36 ret <4 x i16> %tmp2 43 %tmp2 = call <2 x i32> @llvm.ctlz.v2i32(<2 x i32> %tmp1, i1 0) 44 ret <2 x i32> %tmp2 [all …]
|
D | vabs.ll | 7 %tmp2 = call <8 x i8> @llvm.arm.neon.vabs.v8i8(<8 x i8> %tmp1) 8 ret <8 x i8> %tmp2 15 %tmp2 = call <4 x i16> @llvm.arm.neon.vabs.v4i16(<4 x i16> %tmp1) 16 ret <4 x i16> %tmp2 23 %tmp2 = call <2 x i32> @llvm.arm.neon.vabs.v2i32(<2 x i32> %tmp1) 24 ret <2 x i32> %tmp2 31 %tmp2 = call <2 x float> @llvm.fabs.v2f32(<2 x float> %tmp1) 32 ret <2 x float> %tmp2 39 %tmp2 = call <16 x i8> @llvm.arm.neon.vabs.v16i8(<16 x i8> %tmp1) 40 ret <16 x i8> %tmp2 [all …]
|
/external/libunwind/src/dwarf/ |
D | Gexpr.c | 193 unw_word_t operand1 = 0, operand2 = 0, tmp1, tmp2, tmp3, end_addr; in dwarf_eval_expr() local 368 if ((ret = dwarf_readw (as, a, &tmp1, &tmp2, arg)) < 0) in dwarf_eval_expr() 370 push (tmp2); in dwarf_eval_expr() 386 tmp2 = u8; in dwarf_eval_expr() 392 tmp2 = u16; in dwarf_eval_expr() 399 tmp2 = u32; in dwarf_eval_expr() 403 tmp2 >>= 8; in dwarf_eval_expr() 405 tmp2 &= 0xffffff; in dwarf_eval_expr() 414 tmp2 = u64; in dwarf_eval_expr() 418 tmp2 >>= 64 - 8 * operand1; in dwarf_eval_expr() [all …]
|
/external/opencv3/3rdparty/libjpeg/ |
D | jfdctint.c | 160 INT32 tmp0, tmp1, tmp2, tmp3; variable 182 tmp2 = GETJSAMPLE(elemptr[2]) + GETJSAMPLE(elemptr[5]); 187 tmp11 = tmp1 + tmp2; 188 tmp13 = tmp1 - tmp2; 192 tmp2 = GETJSAMPLE(elemptr[2]) - GETJSAMPLE(elemptr[5]); 213 tmp11 = tmp1 + tmp2; 214 tmp12 = tmp0 + tmp2; 222 tmp2 = MULTIPLY(tmp2, FIX_2_053119869); /* c1+c3-c5+c7 */ 237 RIGHT_SHIFT(tmp2 + tmp11 + tmp12, CONST_BITS-PASS1_BITS); 257 tmp2 = dataptr[DCTSIZE*2] + dataptr[DCTSIZE*5]; [all …]
|
/external/llvm/test/Transforms/Reassociate/ |
D | otherops.ll | 7 ; CHECK-NEXT: %tmp2 = mul i32 %arg, 144 8 ; CHECK-NEXT: ret i32 %tmp2 11 %tmp2 = mul i32 %tmp1, 12 12 ret i32 %tmp2 17 ; CHECK-NEXT: %tmp2 = and i32 %arg, 14 18 ; CHECK-NEXT: ret i32 %tmp2 21 %tmp2 = and i32 %tmp1, 14 22 ret i32 %tmp2 27 ; CHECK-NEXT: %tmp2 = or i32 %arg, 14 28 ; CHECK-NEXT: ret i32 %tmp2 [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | arm64-neon-scalar-by-elem-mul.ll | 7 %tmp2 = fmul float %a, %tmp1; 8 ret float %tmp2; 15 %tmp2 = fmul float %tmp1, %a; 16 ret float %tmp2; 24 %tmp2 = fmul float %a, %tmp1; 25 ret float %tmp2; 32 %tmp2 = fmul float %tmp1, %a; 33 ret float %tmp2; 41 %tmp2 = fmul double %a, %tmp1; 42 ret double %tmp2; [all …]
|
D | neon-mla-mls.ll | 7 %tmp2 = add <8 x i8> %C, %tmp1; 8 ret <8 x i8> %tmp2 14 %tmp2 = add <16 x i8> %C, %tmp1; 15 ret <16 x i8> %tmp2 21 %tmp2 = add <4 x i16> %C, %tmp1; 22 ret <4 x i16> %tmp2 28 %tmp2 = add <8 x i16> %C, %tmp1; 29 ret <8 x i16> %tmp2 35 %tmp2 = add <2 x i32> %C, %tmp1; 36 ret <2 x i32> %tmp2 [all …]
|
/external/aac/libFDK/include/arm/ |
D | cplx_mul.h | 110 LONG tmp1,tmp2; in cplxMultDiv2() local 118 : "=&r"(tmp1), "=&r"(tmp2) in cplxMultDiv2() 123 *c_Im = tmp2; in cplxMultDiv2() 135 LONG tmp1, tmp2; in cplxMultDiv2() local 142 : "=&r"(tmp1), "=&r"(tmp2) in cplxMultDiv2() 147 *c_Im = tmp2; in cplxMultDiv2() 159 LONG tmp1, tmp2; in cplxMultAddDiv2() local 166 : "=&r"(tmp1), "=&r"(tmp2) in cplxMultAddDiv2() 171 *c_Im += tmp2; in cplxMultAddDiv2() 184 LONG tmp1, tmp2; in cplxMultDiv2() local [all …]
|
/external/valgrind/none/tests/s390x/ |
D | mul.h | 6 unsigned long tmp2 = m1; \ 12 : "+d" (tmp1), "+d" (tmp2) \ 15 printf(#insn " %16.16lX * %16.16lX = %16.16lX%16.16lX\n", m1, m2, tmp1, tmp2); \ 21 unsigned long tmp2 = m1; \ 27 : "+d" (tmp1), "+d" (tmp2) \ 30 printf(#insn " %16.16lX * %16.16lX = %16.16lX%16.16lX\n", m1, m2, tmp1, tmp2); \ 36 unsigned long tmp2 = m1; \ 42 : "+d" (tmp1), "+d" (tmp2) \ 44 printf(#insn " %16.16lX * %16.16lX = %16.16lX%16.16lX\n", m1, (unsigned long) m2, tmp1, tmp2); \ 50 unsigned long tmp2 = m1; \ [all …]
|
/external/webrtc/webrtc/common_audio/signal_processing/ |
D | resample_by_2_mips.c | 151 int32_t tmp1, tmp2, diff; in WebRtcSpl_DownsampleBy2() local 161 tmp2 = MUL_ACCUM_2(kResampleAllpass2[1], diff, state1); in WebRtcSpl_DownsampleBy2() 163 diff = tmp2 - state3; in WebRtcSpl_DownsampleBy2() 165 state2 = tmp2; in WebRtcSpl_DownsampleBy2() 173 tmp2 = MUL_ACCUM_1(kResampleAllpass1[1], diff, state5); in WebRtcSpl_DownsampleBy2() 175 diff = tmp2 - state7; in WebRtcSpl_DownsampleBy2() 177 state6 = tmp2; in WebRtcSpl_DownsampleBy2() 190 tmp2 = MUL_ACCUM_2(kResampleAllpass2[1], diff, state1); in WebRtcSpl_DownsampleBy2() 192 diff = tmp2 - state3; in WebRtcSpl_DownsampleBy2() 194 state2 = tmp2; in WebRtcSpl_DownsampleBy2() [all …]
|