/external/libvpx/libvpx/vp9/encoder/mips/msa/ |
D | vp9_fdct8x8_msa.c | 18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_fht8x8_msa() local 20 LD_SH8(input, stride, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa() 22 SLLI_4V(in4, in5, in6, in7, 2); in vp9_fht8x8_msa() 26 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_fht8x8_msa() 27 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa() 28 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_fht8x8_msa() 29 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa() 30 VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_fht8x8_msa() 31 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_fht8x8_msa() 34 VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_fht8x8_msa() [all …]
|
D | vp9_fdct4x4_msa.c | 18 v8i16 in0, in1, in2, in3, in4; in vp9_fwht4x4_msa() local 24 in4 = (in0 - in3) >> 1; in vp9_fwht4x4_msa() 25 SUB2(in4, in1, in4, in2, in1, in2); in vp9_fwht4x4_msa() 33 in4 = (in0 - in1) >> 1; in vp9_fwht4x4_msa() 34 SUB2(in4, in2, in4, in3, in2, in3); in vp9_fwht4x4_msa()
|
D | vp9_fdct_msa.h | 18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 35 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \ 38 in4, in3); \ 63 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \
|
D | vp9_fdct16x16_msa.c | 368 v8i16 in0, in1, in2, in3, in4, in5, in6, in7, in8, in9, in10, in11; in postproc_fdct16x8_1d_row() local 371 LD_SH8(temp, 16, in0, in1, in2, in3, in4, in5, in6, in7); in postproc_fdct16x8_1d_row() 374 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in postproc_fdct16x8_1d_row() 375 in0, in1, in2, in3, in4, in5, in6, in7); in postproc_fdct16x8_1d_row() 380 FDCT_POSTPROC_2V_NEG_H(in4, in5); in postproc_fdct16x8_1d_row() 386 BUTTERFLY_16(in0, in1, in2, in3, in4, in5, in6, in7, in postproc_fdct16x8_1d_row() 397 in0, in1, in2, in3, in4, in5, in6, in7); in postproc_fdct16x8_1d_row() 401 TRANSPOSE8x8_SH_SH(tmp4, in4, tmp5, in5, tmp6, in6, tmp7, in7, in postproc_fdct16x8_1d_row() 402 tmp4, in4, tmp5, in5, tmp6, in6, tmp7, in7); in postproc_fdct16x8_1d_row() 404 ST_SH8(tmp4, in4, tmp5, in5, tmp6, in6, tmp7, in7, out, 16); in postproc_fdct16x8_1d_row()
|
/external/libvpx/libvpx/vp9/common/mips/msa/ |
D | vp9_idct8x8_msa.c | 18 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vp9_iht8x8_64_add_msa() local 21 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa() 23 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_iht8x8_64_add_msa() 24 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa() 29 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_iht8x8_64_add_msa() 30 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa() 32 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_iht8x8_64_add_msa() 33 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa() 34 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in vp9_iht8x8_64_add_msa() 35 in0, in1, in2, in3, in4, in5, in6, in7); in vp9_iht8x8_64_add_msa() [all …]
|
/external/libvpx/libvpx/vpx_dsp/mips/ |
D | idct8x8_msa.c | 15 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_64_add_msa() local 18 LD_SH8(input, 8, in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa() 21 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_msa() 22 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa() 24 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_msa() 25 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa() 27 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_msa() 28 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa() 30 VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_msa() 31 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_msa() [all …]
|
D | fwd_txfm_msa.c | 16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x16_1d_column() local 29 in0, in1, in2, in3, in4, in5, in6, in7, in fdct8x16_1d_column() 32 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x16_1d_column() 36 ADD4(in4, in11, in5, in10, in6, in9, in7, in8, tmp4, tmp5, tmp6, tmp7); in fdct8x16_1d_column() 41 SUB4(in4, in11, in5, in10, in6, in9, in7, in8, in11, in10, in9, in8); in fdct8x16_1d_column() 135 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct16x8_1d_row() local 138 LD_SH8(input, 16, in0, in1, in2, in3, in4, in5, in6, in7); in fdct16x8_1d_row() 140 TRANSPOSE8x8_SH_SH(in0, in1, in2, in3, in4, in5, in6, in7, in fdct16x8_1d_row() 141 in0, in1, in2, in3, in4, in5, in6, in7); in fdct16x8_1d_row() 145 ADD4(in4, 1, in5, 1, in6, 1, in7, 1, in4, in5, in6, in7); in fdct16x8_1d_row() [all …]
|
D | fwd_dct32x32_msa.c | 16 v8i16 in0, in1, in2, in3, in4, in5, in6, in7; in fdct8x32_1d_column_load_butterfly() local 23 LD_SH4(input + (28 * src_stride), src_stride, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly() 27 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x32_1d_column_load_butterfly() 30 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, in fdct8x32_1d_column_load_butterfly() 31 step0, step1, step2, step3, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly() 35 ST_SH4(in4, in5, in6, in7, temp_buff + (28 * 8), 8); in fdct8x32_1d_column_load_butterfly() 41 LD_SH4(input + (20 * src_stride), src_stride, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly() 45 SLLI_4V(in4, in5, in6, in7, 2); in fdct8x32_1d_column_load_butterfly() 48 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, in fdct8x32_1d_column_load_butterfly() 49 step0, step1, step2, step3, in4, in5, in6, in7); in fdct8x32_1d_column_load_butterfly() [all …]
|
D | macros_msa.h | 412 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 415 ST_B4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \ 436 #define ST_H8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, pdst, stride) { \ argument 438 ST_H4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \ 593 #define AVER_UB4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 596 AVER_UB2(RTYPE, in4, in5, in6, in7, out2, out3) \ 1057 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 1060 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \ 1108 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 1111 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \ [all …]
|
D | inv_txfm_msa.h | 18 #define VP9_ADST8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 35 ILVRL_H2_SH(in4, in3, vec3_m, vec2_m); \ 38 in4, in3); \ 64 ILVRL_H2_SH(in4, in3, vec1_m, vec0_m); \ 213 #define VP9_IDCT8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 239 VP9_MADD(in0, in4, in2, in6, k1_m, k0_m, k2_m, k3_m, \ 240 in0, in4, in2, in6); \ 241 BUTTERFLY_4(in0, in4, in2, in6, tp0_m, tp1_m, tp2_m, tp3_m); \ 246 #define VP9_IADST8x8_1D(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 265 ILVRL_H2_SH(in5, in4, in_s1, in_s0); \ [all …]
|
D | fwd_txfm_msa.h | 59 #define SRLI_AVE_S_4V_H(in0, in1, in2, in3, in4, in5, in6, in7) { \ argument 63 SRLI_H4_SH(in4, in5, in6, in7, vec4_m, vec5_m, vec6_m, vec7_m, 15); \ 66 AVE_SH4_SH(vec4_m, in4, vec5_m, in5, vec6_m, in6, vec7_m, in7, \ 67 in4, in5, in6, in7); \ 70 #define VP9_FDCT8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 79 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, \ 130 #define FDCT8x16_EVEN(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 138 BUTTERFLY_8(in0, in1, in2, in3, in4, in5, in6, in7, \
|
D | txfm_macros_msa.h | 34 #define DOT_ADD_SUB_SRARI_PCK(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 39 DOTP_SH4_SW(in0, in1, in0, in1, in4, in4, in5, in5, \
|
D | vpx_convolve_msa.h | 109 #define PCKEV_AVG_ST8x4_UB(in1, dst0, in2, dst1, in3, dst2, in4, dst3, \ argument 114 PCKEV_B2_UB(in2, in1, in4, in3, tmp0_m, tmp1_m); \
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_dct_ssse3.c | 54 __m128i in4 = _mm_load_si128((const __m128i *)(input + 4 * stride)); in vp9_fdct8x8_quant_ssse3() local 71 in4 = _mm_slli_epi16(in4, 2); in vp9_fdct8x8_quant_ssse3() 80 in[4] = &in4; in vp9_fdct8x8_quant_ssse3() 96 const __m128i q3 = _mm_add_epi16(in3, in4); in vp9_fdct8x8_quant_ssse3() 97 const __m128i q4 = _mm_sub_epi16(in3, in4); in vp9_fdct8x8_quant_ssse3() 246 in4 = _mm_unpacklo_epi64(tr1_1, tr1_5); in vp9_fdct8x8_quant_ssse3() 269 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in vp9_fdct8x8_quant_ssse3() 277 in4 = _mm_sub_epi16(in4, sign_in4); in vp9_fdct8x8_quant_ssse3() 285 in4 = _mm_srai_epi16(in4, 1); in vp9_fdct8x8_quant_ssse3()
|
D | vp9_dct_sse2.c | 215 __m128i in4 = _mm_load_si128((const __m128i *)(input + 4 * stride)); in vp9_fdct8x8_quant_sse2() local 232 in4 = _mm_slli_epi16(in4, 2); in vp9_fdct8x8_quant_sse2() 241 in[4] = &in4; in vp9_fdct8x8_quant_sse2() 257 const __m128i q3 = _mm_add_epi16(in3, in4); in vp9_fdct8x8_quant_sse2() 258 const __m128i q4 = _mm_sub_epi16(in3, in4); in vp9_fdct8x8_quant_sse2() 413 in4 = _mm_unpacklo_epi64(tr1_1, tr1_5); in vp9_fdct8x8_quant_sse2() 436 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in vp9_fdct8x8_quant_sse2() 444 in4 = _mm_sub_epi16(in4, sign_in4); in vp9_fdct8x8_quant_sse2() 452 in4 = _mm_srai_epi16(in4, 1); in vp9_fdct8x8_quant_sse2() 930 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in fadst8_sse2() local [all …]
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | vp8_macros_msa.h | 378 #define ST_B8(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 382 ST_B4(RTYPE, in4, in5, in6, in7, (pdst) + 4 * stride, stride); \ 577 #define VSHF_B3(RTYPE, in0, in1, in2, in3, in4, in5, mask0, mask1, mask2, \ argument 581 out2 = (RTYPE)__msa_vshf_b((v16i8)mask2, (v16i8)in5, (v16i8)in4); \ 974 #define ILVL_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 978 ILVL_B2(RTYPE, in4, in5, in6, in7, out2, out3); \ 1029 #define ILVR_B4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 1033 ILVR_B2(RTYPE, in4, in5, in6, in7, out2, out3); \ 1056 #define ILVR_H4(RTYPE, in0, in1, in2, in3, in4, in5, in6, in7, \ argument 1060 ILVR_H2(RTYPE, in4, in5, in6, in7, out2, out3); \ [all …]
|
D | postproc_msa.c | 63 #define VP8_TRANSPOSE8x16_UB_UB(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 72 ILVR_B4_SH(in1, in0, in3, in2, in5, in4, in7, in6, \ 78 ILVL_B4_SH(in1, in0, in3, in2, in5, in4, in7, in6, \ 121 #define TRANSPOSE12x16_B(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 129 ILVR_B2_SH(in5, in4, in7, in6, temp0, temp1); \ 144 ILVL_B2_SH(in5, in4, in7, in6, temp0, temp1); \ 145 ILVR_D2_UB(temp6, temp2, temp7, temp3, in4, in6); \ 160 #define VP8_TRANSPOSE12x8_UB_UB(in0, in1, in2, in3, in4, in5, \ argument 168 ILVR_B2_SH(in5, in4, in7, in6, temp0, temp1); \ 174 ILVL_B2_SH(in5, in4, in7, in6, temp6, temp7); \ [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | fwd_txfm_impl_sse2.h | 290 __m128i in4 = _mm_load_si128((const __m128i *)(input + 4 * stride)); in FDCT8x8_2D() local 299 in4 = _mm_slli_epi16(in4, 2); in FDCT8x8_2D() 315 const __m128i q3 = ADD_EPI16(in3, in4); in FDCT8x8_2D() 316 const __m128i q4 = SUB_EPI16(in3, in4); in FDCT8x8_2D() 522 in4 = _mm_unpacklo_epi64(tr1_1, tr1_5); in FDCT8x8_2D() 545 const __m128i sign_in4 = _mm_srai_epi16(in4, 15); in FDCT8x8_2D() 553 in4 = _mm_sub_epi16(in4, sign_in4); in FDCT8x8_2D() 561 in4 = _mm_srai_epi16(in4, 1); in FDCT8x8_2D() 570 store_output(&in4, (output + 4 * 8)); in FDCT8x8_2D()
|
D | inv_txfm_sse2.c | 264 #define TRANSPOSE_8X8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 271 const __m128i tr0_4 = _mm_unpacklo_epi16(in4, in5); \ 273 const __m128i tr0_6 = _mm_unpackhi_epi16(in4, in5); \ 380 #define IDCT8(in0, in1, in2, in3, in4, in5, in6, in7, \ argument 397 const __m128i lo_04 = _mm_unpacklo_epi16(in0, in4); \ 398 const __m128i hi_04 = _mm_unpackhi_epi16(in0, in4); \ 465 __m128i in0, in1, in2, in3, in4, in5, in6, in7; in vpx_idct8x8_64_add_sse2() local 476 in4 = _mm_load_si128((const __m128i *)(input + 8 * 4)); in vpx_idct8x8_64_add_sse2() 484 TRANSPOSE_8X8(in0, in1, in2, in3, in4, in5, in6, in7, in vpx_idct8x8_64_add_sse2() 485 in0, in1, in2, in3, in4, in5, in6, in7); in vpx_idct8x8_64_add_sse2() [all …]
|
/external/openssh/ |
D | audit-bsm.c | 141 struct sockaddr_in *in4; in aug_get_machine() local 153 in4 = (struct sockaddr_in *)ai->ai_addr; in aug_get_machine() 155 memcpy(addr, &in4->sin_addr, sizeof(struct in_addr)); in aug_get_machine()
|
D | addrmatch.c | 84 struct sockaddr_in *in4 = (struct sockaddr_in *)sa; in addr_sa_to_xaddr() local 91 if (slen < (socklen_t)sizeof(*in4)) in addr_sa_to_xaddr() 94 memcpy(&xa->v4, &in4->sin_addr, sizeof(xa->v4)); in addr_sa_to_xaddr()
|
/external/llvm/test/Transforms/LoopUnswitch/ |
D | 2007-08-01-LCSSA.ll | 34 %iftmp.37.0.in4 = icmp eq i32 %tmp138, 0 ; <i1> [#uses=1] 35 br i1 %iftmp.37.0.in4, label %bb250, label %bb166
|
/external/boringssl/src/crypto/aes/asm/ |
D | aesni-x86_64.pl | 1160 my ($in0,$in1,$in2,$in3,$in4,$in5)=map("%xmm$_",(10..15)); 1481 movdqu 0x40($inp),$in4 1484 pxor $rndkey0,$in4 1507 aesenclast $in4,$inout4 1509 movdqa 0x30(%rsp),$in4 1523 movdqa $in4,$inout3 2719 my ($iv,$in0,$in1,$in2,$in3,$in4)=map("%xmm$_",(10..15)); 2838 movdqa $inout4,$in4 2926 pxor $rndkey0,$in4 2944 aesdeclast $in4,$inout5 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | tailcall-stackalign.ll | 10 define fastcc i32 @tailcaller(i32 %in1, i32 %in2, i32 %in3, i32 %in4) {
|
/external/compiler-rt/test/dfsan/ |
D | custom.cc | 535 struct in_addr in4; in test_inet_pton() local 536 int ret4 = inet_pton(AF_INET, addr4, &in4); in test_inet_pton() 538 ASSERT_READ_LABEL(&in4, sizeof(in4), i_label); in test_inet_pton() 539 assert(in4.s_addr == 0x0100007f); in test_inet_pton()
|