/external/webrtc/webrtc/common_audio/signal_processing/ |
D | resample_by_2_internal.c | 34 int32_t tmp0, tmp1, diff; in WebRtcSpl_DownBy2IntToShort() local 46 tmp1 = state[0] + diff * kResampleAllpass[1][0]; in WebRtcSpl_DownBy2IntToShort() 48 diff = tmp1 - state[2]; in WebRtcSpl_DownBy2IntToShort() 54 state[1] = tmp1; in WebRtcSpl_DownBy2IntToShort() 76 tmp1 = state[4] + diff * kResampleAllpass[0][0]; in WebRtcSpl_DownBy2IntToShort() 78 diff = tmp1 - state[6]; in WebRtcSpl_DownBy2IntToShort() 84 state[5] = tmp1; in WebRtcSpl_DownBy2IntToShort() 104 tmp1 = (in[(i << 1) + 2] + in[(i << 1) + 3]) >> 15; in WebRtcSpl_DownBy2IntToShort() 110 if (tmp1 > (int32_t)0x00007FFF) in WebRtcSpl_DownBy2IntToShort() 111 tmp1 = 0x00007FFF; in WebRtcSpl_DownBy2IntToShort() [all …]
|
D | resample_by_2_mips.c | 151 int32_t tmp1, tmp2, diff; in WebRtcSpl_DownsampleBy2() local 158 tmp1 = MUL_ACCUM_1(kResampleAllpass2[0], diff, state0); in WebRtcSpl_DownsampleBy2() 160 diff = tmp1 - state2; in WebRtcSpl_DownsampleBy2() 162 state1 = tmp1; in WebRtcSpl_DownsampleBy2() 170 tmp1 = MUL_ACCUM_1(kResampleAllpass1[0], diff, state4); in WebRtcSpl_DownsampleBy2() 172 diff = tmp1 - state6; in WebRtcSpl_DownsampleBy2() 174 state5 = tmp1; in WebRtcSpl_DownsampleBy2() 187 tmp1 = MUL_ACCUM_1(kResampleAllpass2[0], diff, state0); in WebRtcSpl_DownsampleBy2() 189 diff = tmp1 - state2; in WebRtcSpl_DownsampleBy2() 191 state1 = tmp1; in WebRtcSpl_DownsampleBy2() [all …]
|
/external/pdfium/core/src/fxcodec/jbig2/ |
D | JBig2_Image.cpp | 244 tmp1 = 0, tmp2 = 0, maskL = 0, maskR = 0, maskM = 0; in composeTo_opt2() local 297 tmp1 = JBIG2_GETDWORD(lineSrc) << shift; in composeTo_opt2() 301 tmp = (tmp2 & ~maskM) | ((tmp1 | tmp2) & maskM); in composeTo_opt2() 304 tmp = (tmp2 & ~maskM) | ((tmp1 & tmp2) & maskM); in composeTo_opt2() 307 tmp = (tmp2 & ~maskM) | ((tmp1 ^ tmp2) & maskM); in composeTo_opt2() 310 tmp = (tmp2 & ~maskM) | ((~(tmp1 ^ tmp2)) & maskM); in composeTo_opt2() 313 tmp = (tmp2 & ~maskM) | (tmp1 & maskM); in composeTo_opt2() 326 tmp1 = JBIG2_GETDWORD(lineSrc) >> shift; in composeTo_opt2() 330 tmp = (tmp2 & ~maskM) | ((tmp1 | tmp2) & maskM); in composeTo_opt2() 333 tmp = (tmp2 & ~maskM) | ((tmp1 & tmp2) & maskM); in composeTo_opt2() [all …]
|
/external/libunwind/src/dwarf/ |
D | Gexpr.c | 193 unw_word_t operand1 = 0, operand2 = 0, tmp1, tmp2, tmp3, end_addr; in dwarf_eval_expr() local 296 &tmp1)) < 0) in dwarf_eval_expr() 298 push (tmp1 + operand1); in dwarf_eval_expr() 305 dwarf_to_unw_regnum (operand1), &tmp1)) < 0) in dwarf_eval_expr() 307 push (tmp1 + operand2); in dwarf_eval_expr() 367 tmp1 = pop (); in dwarf_eval_expr() 368 if ((ret = dwarf_readw (as, a, &tmp1, &tmp2, arg)) < 0) in dwarf_eval_expr() 375 tmp1 = pop (); in dwarf_eval_expr() 384 if ((ret = dwarf_readu8 (as, a, &tmp1, &u8, arg)) < 0) in dwarf_eval_expr() 390 if ((ret = dwarf_readu16 (as, a, &tmp1, &u16, arg)) < 0) in dwarf_eval_expr() [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | vst4.ll | 7 %tmp1 = load <8 x i8>, <8 x i8>* %B 8 …d @llvm.arm.neon.vst4.p0i8.v8i8(i8* %A, <8 x i8> %tmp1, <8 x i8> %tmp1, <8 x i8> %tmp1, <8 x i8> %… 17 %tmp1 = load <8 x i8>, <8 x i8>* %B 18 …d @llvm.arm.neon.vst4.p0i8.v8i8(i8* %A, <8 x i8> %tmp1, <8 x i8> %tmp1, <8 x i8> %tmp1, <8 x i8> %… 29 %tmp1 = load <4 x i16>, <4 x i16>* %B 30 …m.arm.neon.vst4.p0i8.v4i16(i8* %tmp0, <4 x i16> %tmp1, <4 x i16> %tmp1, <4 x i16> %tmp1, <4 x i16>… 39 %tmp1 = load <2 x i32>, <2 x i32>* %B 40 …m.arm.neon.vst4.p0i8.v2i32(i8* %tmp0, <2 x i32> %tmp1, <2 x i32> %tmp1, <2 x i32> %tmp1, <2 x i32>… 48 %tmp1 = load <2 x float>, <2 x float>* %B 49 ….neon.vst4.p0i8.v2f32(i8* %tmp0, <2 x float> %tmp1, <2 x float> %tmp1, <2 x float> %tmp1, <2 x flo… [all …]
|
D | vstlane.ll | 7 %tmp1 = load <8 x i8>, <8 x i8>* %B 8 %tmp2 = extractelement <8 x i8> %tmp1, i32 3 18 %tmp1 = load <8 x i8>, <8 x i8>* %B 19 %tmp2 = extractelement <8 x i8> %tmp1, i32 3 30 %tmp1 = load <4 x i16>, <4 x i16>* %B 31 %tmp2 = extractelement <4 x i16> %tmp1, i32 2 40 %tmp1 = load <2 x i32>, <2 x i32>* %B 41 %tmp2 = extractelement <2 x i32> %tmp1, i32 1 49 %tmp1 = load <2 x float>, <2 x float>* %B 50 %tmp2 = extractelement <2 x float> %tmp1, i32 1 [all …]
|
D | vshift.ll | 6 %tmp1 = load <8 x i8>, <8 x i8>* %A 8 %tmp3 = shl <8 x i8> %tmp1, %tmp2 15 %tmp1 = load <4 x i16>, <4 x i16>* %A 17 %tmp3 = shl <4 x i16> %tmp1, %tmp2 24 %tmp1 = load <2 x i32>, <2 x i32>* %A 26 %tmp3 = shl <2 x i32> %tmp1, %tmp2 33 %tmp1 = load <1 x i64>, <1 x i64>* %A 35 %tmp3 = shl <1 x i64> %tmp1, %tmp2 42 %tmp1 = load <8 x i8>, <8 x i8>* %A 43 %tmp2 = shl <8 x i8> %tmp1, < i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7, i8 7 > [all …]
|
D | vbits.ll | 6 %tmp1 = load <8 x i8>, <8 x i8>* %A 8 %tmp3 = and <8 x i8> %tmp1, %tmp2 15 %tmp1 = load <4 x i16>, <4 x i16>* %A 17 %tmp3 = and <4 x i16> %tmp1, %tmp2 24 %tmp1 = load <2 x i32>, <2 x i32>* %A 26 %tmp3 = and <2 x i32> %tmp1, %tmp2 33 %tmp1 = load <1 x i64>, <1 x i64>* %A 35 %tmp3 = and <1 x i64> %tmp1, %tmp2 42 %tmp1 = load <16 x i8>, <16 x i8>* %A 44 %tmp3 = and <16 x i8> %tmp1, %tmp2 [all …]
|
D | vst3.ll | 8 %tmp1 = load <8 x i8>, <8 x i8>* %B 9 …call void @llvm.arm.neon.vst3.p0i8.v8i8(i8* %A, <8 x i8> %tmp1, <8 x i8> %tmp1, <8 x i8> %tmp1, i3… 17 %tmp1 = load <4 x i16>, <4 x i16>* %B 18 …oid @llvm.arm.neon.vst3.p0i8.v4i16(i8* %tmp0, <4 x i16> %tmp1, <4 x i16> %tmp1, <4 x i16> %tmp1, i… 26 %tmp1 = load <2 x i32>, <2 x i32>* %B 27 …oid @llvm.arm.neon.vst3.p0i8.v2i32(i8* %tmp0, <2 x i32> %tmp1, <2 x i32> %tmp1, <2 x i32> %tmp1, i… 37 %tmp1 = load <2 x i32>, <2 x i32>* %B 38 …oid @llvm.arm.neon.vst3.p0i8.v2i32(i8* %tmp0, <2 x i32> %tmp1, <2 x i32> %tmp1, <2 x i32> %tmp1, i… 48 %tmp1 = load <2 x float>, <2 x float>* %B 49 …@llvm.arm.neon.vst3.p0i8.v2f32(i8* %tmp0, <2 x float> %tmp1, <2 x float> %tmp1, <2 x float> %tmp1,… [all …]
|
D | uxtb.ll | 5 %tmp1 = and i32 %x, 16711935 ; <i32> [#uses=1] 6 ret i32 %tmp1 10 %tmp1 = lshr i32 %x, 8 ; <i32> [#uses=1] 11 %tmp2 = and i32 %tmp1, 16711935 ; <i32> [#uses=1] 16 %tmp1 = lshr i32 %x, 8 ; <i32> [#uses=1] 17 %tmp2 = and i32 %tmp1, 16711935 ; <i32> [#uses=1] 22 %tmp1 = lshr i32 %x, 8 ; <i32> [#uses=1] 23 %tmp6 = and i32 %tmp1, 16711935 ; <i32> [#uses=1] 28 %tmp1 = lshr i32 %x, 8 ; <i32> [#uses=1] 29 %tmp2 = and i32 %tmp1, 16711935 ; <i32> [#uses=1] [all …]
|
D | vneg.ll | 6 %tmp1 = load <8 x i8>, <8 x i8>* %A 7 %tmp2 = sub <8 x i8> zeroinitializer, %tmp1 14 %tmp1 = load <4 x i16>, <4 x i16>* %A 15 %tmp2 = sub <4 x i16> zeroinitializer, %tmp1 22 %tmp1 = load <2 x i32>, <2 x i32>* %A 23 %tmp2 = sub <2 x i32> zeroinitializer, %tmp1 30 %tmp1 = load <2 x float>, <2 x float>* %A 31 %tmp2 = fsub <2 x float> < float -0.000000e+00, float -0.000000e+00 >, %tmp1 38 %tmp1 = load <16 x i8>, <16 x i8>* %A 39 %tmp2 = sub <16 x i8> zeroinitializer, %tmp1 [all …]
|
D | vst2.ll | 7 %tmp1 = load <8 x i8>, <8 x i8>* %B 8 call void @llvm.arm.neon.vst2.p0i8.v8i8(i8* %A, <8 x i8> %tmp1, <8 x i8> %tmp1, i32 8) 17 %tmp1 = load <8 x i8>, <8 x i8>* %B 18 call void @llvm.arm.neon.vst2.p0i8.v8i8(i8* %A, <8 x i8> %tmp1, <8 x i8> %tmp1, i32 4) 29 %tmp1 = load <4 x i16>, <4 x i16>* %B 30 call void @llvm.arm.neon.vst2.p0i8.v4i16(i8* %tmp0, <4 x i16> %tmp1, <4 x i16> %tmp1, i32 32) 38 %tmp1 = load <2 x i32>, <2 x i32>* %B 39 call void @llvm.arm.neon.vst2.p0i8.v2i32(i8* %tmp0, <2 x i32> %tmp1, <2 x i32> %tmp1, i32 1) 47 %tmp1 = load <2 x float>, <2 x float>* %B 48 call void @llvm.arm.neon.vst2.p0i8.v2f32(i8* %tmp0, <2 x float> %tmp1, <2 x float> %tmp1, i32 1) [all …]
|
/external/llvm/test/CodeGen/AArch64/ |
D | neon-bitwise-instructions.ll | 6 %tmp1 = and <8 x i8> %a, %b; 7 ret <8 x i8> %tmp1 13 %tmp1 = and <16 x i8> %a, %b; 14 ret <16 x i8> %tmp1 21 %tmp1 = or <8 x i8> %a, %b; 22 ret <8 x i8> %tmp1 28 %tmp1 = or <16 x i8> %a, %b; 29 ret <16 x i8> %tmp1 36 %tmp1 = xor <8 x i8> %a, %b; 37 ret <8 x i8> %tmp1 [all …]
|
D | neon-scalar-by-elem-fma.ll | 9 %tmp1 = extractelement <4 x float> %v, i32 3 10 %tmp2 = call float @llvm.fma.f32(float %b, float %tmp1, float %a) 17 %tmp1 = extractelement <4 x float> %v, i32 3 18 %tmp2 = call float @llvm.fma.f32(float %tmp1, float %a, float %a) 25 %tmp1 = extractelement <2 x float> %v, i32 1 26 %tmp2 = call float @llvm.fma.f32(float %b, float %tmp1, float %a) 33 %tmp1 = extractelement <1 x double> %v, i32 0 34 %tmp2 = call double @llvm.fma.f64(double %b, double %tmp1, double %a) 41 %tmp1 = extractelement <2 x double> %v, i32 1 42 %tmp2 = call double @llvm.fma.f64(double %b, double %tmp1, double %a) [all …]
|
D | arm64-neon-scalar-by-elem-mul.ll | 6 %tmp1 = extractelement <2 x float> %v, i32 1 7 %tmp2 = fmul float %a, %tmp1; 14 %tmp1 = extractelement <2 x float> %v, i32 1 15 %tmp2 = fmul float %tmp1, %a; 23 %tmp1 = extractelement <4 x float> %v, i32 3 24 %tmp2 = fmul float %a, %tmp1; 31 %tmp1 = extractelement <4 x float> %v, i32 3 32 %tmp2 = fmul float %tmp1, %a; 40 %tmp1 = extractelement <1 x double> %v, i32 0 41 %tmp2 = fmul double %a, %tmp1; [all …]
|
D | neon-mla-mls.ll | 6 %tmp1 = mul <8 x i8> %A, %B; 7 %tmp2 = add <8 x i8> %C, %tmp1; 13 %tmp1 = mul <16 x i8> %A, %B; 14 %tmp2 = add <16 x i8> %C, %tmp1; 20 %tmp1 = mul <4 x i16> %A, %B; 21 %tmp2 = add <4 x i16> %C, %tmp1; 27 %tmp1 = mul <8 x i16> %A, %B; 28 %tmp2 = add <8 x i16> %C, %tmp1; 34 %tmp1 = mul <2 x i32> %A, %B; 35 %tmp2 = add <2 x i32> %C, %tmp1; [all …]
|
/external/webrtc/webrtc/modules/audio_coding/codecs/isac/fix/source/ |
D | filterbanks_neon.c | 45 int16x4_t tmp1, tmp2; in WebRtcIsacfix_AllpassFilter2FixDec16Neon() local 52 tmp1 = vshrn_n_s32(a, 16); in WebRtcIsacfix_AllpassFilter2FixDec16Neon() 55 statev = vqdmlsl_s16(vshll_n_s16(datav, 16), tmp1, factorv); in WebRtcIsacfix_AllpassFilter2FixDec16Neon() 62 tmp1 = vld1_lane_s16(data_ch1 + 1, tmp1, 1); in WebRtcIsacfix_AllpassFilter2FixDec16Neon() 63 tmp1 = vld1_lane_s16(data_ch2 + 1, tmp1, 3); in WebRtcIsacfix_AllpassFilter2FixDec16Neon() 64 datav = vrev32_s16(tmp1); in WebRtcIsacfix_AllpassFilter2FixDec16Neon() 69 tmp1 = vshrn_n_s32(a, 16); in WebRtcIsacfix_AllpassFilter2FixDec16Neon() 71 vst1_lane_s16(data_ch1 + n, tmp1, 1); in WebRtcIsacfix_AllpassFilter2FixDec16Neon() 72 vst1_lane_s16(data_ch2 + n, tmp1, 3); in WebRtcIsacfix_AllpassFilter2FixDec16Neon() 76 statev = vqdmlsl_s16(vshll_n_s16(datav, 16), tmp1, factorv); in WebRtcIsacfix_AllpassFilter2FixDec16Neon() [all …]
|
/external/boringssl/src/crypto/md5/asm/ |
D | md5-586.pl | 20 $tmp1="edi"; 47 &mov($tmp1,$C) if $pos < 0; 53 &xor($tmp1,$d); # F function - part 2 55 &and($tmp1,$b); # F function - part 3 58 &xor($tmp1,$d); # F function - part 4 60 &add($a,$tmp1); 61 &mov($tmp1,&Np($c)) if $pos < 1; # next tmp1 for R0 62 &mov($tmp1,&Np($c)) if $pos == 1; # next tmp1 for R1 79 &xor($tmp1,$b); # G function - part 2 80 &and($tmp1,$d); # G function - part 3 [all …]
|
/external/opencv3/3rdparty/libjpeg/ |
D | jfdctint.c | 160 INT32 tmp0, tmp1, tmp2, tmp3; variable 181 tmp1 = GETJSAMPLE(elemptr[1]) + GETJSAMPLE(elemptr[6]); 187 tmp11 = tmp1 + tmp2; 188 tmp13 = tmp1 - tmp2; 191 tmp1 = GETJSAMPLE(elemptr[1]) - GETJSAMPLE(elemptr[6]); 213 tmp11 = tmp1 + tmp2; 215 tmp13 = tmp1 + tmp3; 221 tmp1 = MULTIPLY(tmp1, FIX_3_072711026); /* c1+c3+c5-c7 */ 235 RIGHT_SHIFT(tmp1 + tmp11 + tmp13, CONST_BITS-PASS1_BITS); 256 tmp1 = dataptr[DCTSIZE*1] + dataptr[DCTSIZE*6]; [all …]
|
/external/llvm/test/CodeGen/PowerPC/ |
D | bswap-load-store.ll | 8 %tmp1 = getelementptr i8, i8* %ptr, i32 %off ; <i8*> [#uses=1] 9 %tmp1.upgrd.1 = bitcast i8* %tmp1 to i32* ; <i32*> [#uses=1] 11 store i32 %tmp13, i32* %tmp1.upgrd.1 16 %tmp1 = getelementptr i8, i8* %ptr, i32 %off ; <i8*> [#uses=1] 17 %tmp1.upgrd.2 = bitcast i8* %tmp1 to i32* ; <i32*> [#uses=1] 18 %tmp = load i32, i32* %tmp1.upgrd.2 ; <i32> [#uses=1] 24 %tmp1 = getelementptr i8, i8* %ptr, i32 %off ; <i8*> [#uses=1] 25 %tmp1.upgrd.3 = bitcast i8* %tmp1 to i16* ; <i16*> [#uses=1] 27 store i16 %tmp5, i16* %tmp1.upgrd.3 32 %tmp1 = getelementptr i8, i8* %ptr, i32 %off ; <i8*> [#uses=1] [all …]
|
/external/wpa_supplicant_8/src/crypto/ |
D | milenage.c | 39 u8 tmp1[16], tmp2[16], tmp3[16]; in milenage_f1() local 44 tmp1[i] = _rand[i] ^ opc[i]; in milenage_f1() 45 if (aes_128_encrypt_block(k, tmp1, tmp1)) in milenage_f1() 60 tmp3[i] ^= tmp1[i]; in milenage_f1() 64 if (aes_128_encrypt_block(k, tmp3, tmp1)) in milenage_f1() 67 tmp1[i] ^= opc[i]; in milenage_f1() 69 os_memcpy(mac_a, tmp1, 8); /* f1 */ in milenage_f1() 71 os_memcpy(mac_s, tmp1 + 8, 8); /* f1* */ in milenage_f1() 91 u8 tmp1[16], tmp2[16], tmp3[16]; in milenage_f2345() local 96 tmp1[i] = _rand[i] ^ opc[i]; in milenage_f2345() [all …]
|
/external/aac/libFDK/include/arm/ |
D | cplx_mul.h | 110 LONG tmp1,tmp2; in cplxMultDiv2() local 118 : "=&r"(tmp1), "=&r"(tmp2) in cplxMultDiv2() 122 *c_Re = tmp1; in cplxMultDiv2() 135 LONG tmp1, tmp2; in cplxMultDiv2() local 142 : "=&r"(tmp1), "=&r"(tmp2) in cplxMultDiv2() 146 *c_Re = tmp1; in cplxMultDiv2() 159 LONG tmp1, tmp2; in cplxMultAddDiv2() local 166 : "=&r"(tmp1), "=&r"(tmp2) in cplxMultAddDiv2() 170 *c_Re += tmp1; in cplxMultAddDiv2() 184 LONG tmp1, tmp2; in cplxMultDiv2() local [all …]
|
/external/valgrind/none/tests/s390x/ |
D | mul.h | 5 unsigned long tmp1 = m1; \ 12 : "+d" (tmp1), "+d" (tmp2) \ 15 printf(#insn " %16.16lX * %16.16lX = %16.16lX%16.16lX\n", m1, m2, tmp1, tmp2); \ 20 unsigned long tmp1 = m1; \ 27 : "+d" (tmp1), "+d" (tmp2) \ 30 printf(#insn " %16.16lX * %16.16lX = %16.16lX%16.16lX\n", m1, m2, tmp1, tmp2); \ 35 unsigned long tmp1 = m1; \ 42 : "+d" (tmp1), "+d" (tmp2) \ 44 printf(#insn " %16.16lX * %16.16lX = %16.16lX%16.16lX\n", m1, (unsigned long) m2, tmp1, tmp2); \ 49 unsigned long tmp1 = m1; \ [all …]
|
/external/libopus/silk/fixed/ |
D | burg_modified_FIX.c | 58 opus_int32 C0, num, nrg, rc_Q31, invGain_Q30, Atmp_QA, Atmp1, tmp1, tmp2, x1, x2; in silk_burg_modified() local 130 …tmp1 = silk_LSHIFT32( (opus_int32)x_ptr[ n ], QA - 16 ); /* Q(QA-16… in silk_burg_modified() 136 … tmp1 = silk_SMLAWB( tmp1, Atmp_QA, x_ptr[ n - k - 1 ] ); /* Q(QA-16) */ in silk_burg_modified() 139 …tmp1 = silk_LSHIFT32( -tmp1, 32 - QA - rshifts ); /* Q(16-rs… in silk_burg_modified() 142 …CAf[ k ] = silk_SMLAWB( CAf[ k ], tmp1, x_ptr[ n - k ] ); /* Q( -rshift … in silk_burg_modified() 151 … tmp1 = silk_LSHIFT32( (opus_int32)x_ptr[ n ], 17 ); /* Q17 */ in silk_burg_modified() 157 … tmp1 = silk_MLA( tmp1, x_ptr[ n - k - 1 ], Atmp1 ); /* Q17 */ in silk_burg_modified() 160 … tmp1 = -tmp1; /* Q17 */ in silk_burg_modified() 163 CAf[ k ] = silk_SMLAWW( CAf[ k ], tmp1, in silk_burg_modified() 172 …tmp1 = C_first_row[ n ]; /*… in silk_burg_modified() [all …]
|
/external/boringssl/src/crypto/sha/asm/ |
D | sha1-586.pl | 143 $tmp1="ebp"; 158 if ($n==0) { &mov($tmp1,$a); } 159 else { &mov($a,$tmp1); } 160 &rotl($tmp1,5); # tmp1=ROTATE(a,5) 162 &add($tmp1,$e); # tmp1+=e; 169 &lea($tmp1,&DWP(0x5a827999,$tmp1,$e)); # tmp1+=K_00_19+xi 172 &add($f,$tmp1); } # f+=tmp1 173 else { &add($tmp1,$f); } # f becomes a in next round 174 &mov($tmp1,$a) if ($alt && $n==15); 186 &and($tmp1,$c); # tmp1 to hold F_00_19(b,c,d), b&=c^d [all …]
|