/external/libvpx/libvpx/vpx_dsp/mips/ |
D | fwd_dct32x32_msa.c | 62 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in fdct8x32_1d_column_even_store() local 72 BUTTERFLY_8(in4, in5, in6, in7, in8, in9, in10, in11, vec4, vec5, vec6, vec7, in fdct8x32_1d_column_even_store() 76 ADD4(vec0, vec7, vec1, vec6, vec2, vec5, vec3, vec4, in0, in1, in2, in3); in fdct8x32_1d_column_even_store() 88 SUB4(vec0, vec7, vec1, vec6, vec2, vec5, vec3, vec4, vec7, vec6, vec5, vec4); in fdct8x32_1d_column_even_store() 89 DOTP_CONST_PAIR(vec6, vec5, cospi_16_64, cospi_16_64, vec5, vec6); in fdct8x32_1d_column_even_store() 90 ADD2(vec4, vec5, vec7, vec6, vec0, vec1); in fdct8x32_1d_column_even_store() 96 SUB2(vec4, vec5, vec7, vec6, vec4, vec7); in fdct8x32_1d_column_even_store() 102 DOTP_CONST_PAIR(in13, in10, cospi_16_64, cospi_16_64, vec2, vec5); in fdct8x32_1d_column_even_store() 104 ADD4(in8, vec3, in9, vec2, in14, vec5, in15, vec4, in0, vec1, vec6, in2); in fdct8x32_1d_column_even_store() 118 SUB2(in9, vec2, in14, vec5, vec2, vec5); in fdct8x32_1d_column_even_store() [all …]
|
D | vpx_convolve8_avg_horiz_msa.c | 151 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_8t_and_aver_dst_16w_msa() local 172 VSHF_B4_SH(src1, src1, mask0, mask1, mask2, mask3, vec1, vec5, vec9, vec13); in common_hz_8t_and_aver_dst_16w_msa() 181 DPADD_SB4_SH(vec4, vec5, vec6, vec7, filt1, filt1, filt1, filt1, vec0, vec1, in common_hz_8t_and_aver_dst_16w_msa() 205 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_8t_and_aver_dst_32w_msa() local 228 VSHF_B4_SH(src1, src1, mask0, mask1, mask2, mask3, vec1, vec5, vec9, vec13); in common_hz_8t_and_aver_dst_32w_msa() 237 DPADD_SB4_SH(vec4, vec5, vec6, vec7, filt1, filt1, filt1, filt1, vec0, vec1, in common_hz_8t_and_aver_dst_32w_msa() 260 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_8t_and_aver_dst_64w_msa() local 284 VSHF_B4_SH(src1, src1, mask0, mask1, mask2, mask3, vec1, vec5, vec9, in common_hz_8t_and_aver_dst_64w_msa() 294 DPADD_SB4_SH(vec4, vec5, vec6, vec7, filt1, filt1, filt1, filt1, vec0, in common_hz_8t_and_aver_dst_64w_msa() 346 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_and_aver_dst_4x8_msa() local [all …]
|
D | vpx_convolve8_avg_vert_msa.c | 362 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_and_aver_dst_8x8mult_msa() local 385 ILVR_B4_UB(src5, src4, src6, src5, src7, src6, src8, src7, vec4, vec5, vec6, in common_vt_2t_and_aver_dst_8x8mult_msa() 393 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, tmp0, tmp1, in common_vt_2t_and_aver_dst_8x8mult_msa() 421 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_vt_2t_and_aver_dst_16w_msa() local 444 ILVL_B2_UB(src3, src2, src4, src3, vec5, vec7); in common_vt_2t_and_aver_dst_16w_msa() 450 DOTP_UB2_UH(vec4, vec5, filt0, filt0, tmp0, tmp1); in common_vt_2t_and_aver_dst_16w_msa() 471 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_and_aver_dst_32w_msa() local 500 ILVL_B2_UB(src3, src2, src4, src3, vec5, vec7); in common_vt_2t_and_aver_dst_32w_msa() 501 DOTP_UB2_UH(vec4, vec5, filt0, filt0, tmp0, tmp1); in common_vt_2t_and_aver_dst_32w_msa() 520 ILVL_B2_UB(src8, src7, src9, src8, vec5, vec7); in common_vt_2t_and_aver_dst_32w_msa() [all …]
|
D | vpx_convolve8_horiz_msa.c | 344 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_4x8_msa() local 355 DOTP_UB4_UH(vec0, vec1, vec2, vec3, filt0, filt0, filt0, filt0, vec4, vec5, in common_hz_2t_4x8_msa() 357 SRARI_H4_UH(vec4, vec5, vec6, vec7, FILTER_BITS); in common_hz_2t_4x8_msa() 358 PCKEV_B4_SB(vec4, vec4, vec5, vec5, vec6, vec6, vec7, vec7, res0, res1, res2, in common_hz_2t_4x8_msa() 476 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_16w_msa() local 493 VSHF_B2_UB(src4, src4, src5, src5, mask, mask, vec4, vec5); in common_hz_2t_16w_msa() 497 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 517 VSHF_B2_UB(src4, src4, src5, src5, mask, mask, vec4, vec5); in common_hz_2t_16w_msa() 521 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 541 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_32w_msa() local [all …]
|
D | vpx_convolve8_vert_msa.c | 394 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_8x8mult_msa() local 412 ILVR_B4_UB(src5, src4, src6, src5, src7, src6, src8, src7, vec4, vec5, vec6, in common_vt_2t_8x8mult_msa() 421 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, tmp0, tmp1, in common_vt_2t_8x8mult_msa() 447 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_16w_msa() local 470 ILVL_B2_UB(src3, src2, src4, src3, vec5, vec7); in common_vt_2t_16w_msa() 476 DOTP_UB2_UH(vec4, vec5, filt0, filt0, tmp0, tmp1); in common_vt_2t_16w_msa() 495 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_32w_msa() local 523 ILVL_B2_UB(src3, src2, src4, src3, vec5, vec7); in common_vt_2t_32w_msa() 524 DOTP_UB2_UH(vec4, vec5, filt0, filt0, tmp0, tmp1); in common_vt_2t_32w_msa() 543 ILVL_B2_UB(src8, src7, src9, src8, vec5, vec7); in common_vt_2t_32w_msa() [all …]
|
D | fwd_txfm_msa.c | 35 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, cnst0, cnst1, cnst4, cnst5; in fdct8x16_1d_column() local 62 ILVR_H2_SH(in10, in13, in11, in12, vec3, vec5); in fdct8x16_1d_column() 70 stp24 = DOT_SHIFT_RIGHT_PCK_H(vec4, vec5, cnst4); in fdct8x16_1d_column() 71 stp23 = DOT_SHIFT_RIGHT_PCK_H(vec4, vec5, cnst5); in fdct8x16_1d_column() 77 ILVR_H2_SH(stp36, stp31, stp35, stp32, vec3, vec5); in fdct8x16_1d_column() 114 stp25 = DOT_SHIFT_RIGHT_PCK_H(vec4, vec5, cnst1); in fdct8x16_1d_column() 118 stp22 = DOT_SHIFT_RIGHT_PCK_H(vec4, vec5, cnst1); in fdct8x16_1d_column()
|
D | loopfilter_8_msa.c | 243 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in vpx_lpf_vertical_8_dual_msa() local 280 ILVRL_H2_SH(vec1, vec0, vec4, vec5); in vpx_lpf_vertical_8_dual_msa() 285 ST4x8_UB(vec4, vec5, src, pitch); in vpx_lpf_vertical_8_dual_msa() 318 ILVRL_B2_SH(q2, q1, vec2, vec5); in vpx_lpf_vertical_8_dual_msa() 328 ST2x4_UB(vec5, 0, src + 4, pitch); in vpx_lpf_vertical_8_dual_msa() 331 ST2x4_UB(vec5, 4, src + 4, pitch); in vpx_lpf_vertical_8_dual_msa()
|
/external/eigen/unsupported/test/ |
D | cxx11_tensor_simple.cpp | 96 Tensor<int, 1> vec5(vec1); in test_1d() local 98 VERIFY_IS_EQUAL((vec5(0)), 4); in test_1d() 99 VERIFY_IS_EQUAL((vec5(1)), 8); in test_1d() 100 VERIFY_IS_EQUAL((vec5(2)), 15); in test_1d() 101 VERIFY_IS_EQUAL((vec5(3)), 16); in test_1d() 102 VERIFY_IS_EQUAL((vec5(4)), 23); in test_1d() 103 VERIFY_IS_EQUAL((vec5(5)), 42); in test_1d() 105 VERIFY_IS_EQUAL((vec5.data()[0]), 4); in test_1d() 106 VERIFY_IS_EQUAL((vec5.data()[1]), 8); in test_1d() 107 VERIFY_IS_EQUAL((vec5.data()[2]), 15); in test_1d() [all …]
|
D | cxx11_tensor_expr.cpp | 36 TensorMap<Tensor<float, 1, RowMajor>> vec5(data5, 6); in test_1d() local 37 vec5 = vec2.cube(); in test_1d() 53 VERIFY_IS_APPROX(vec5(0), 0.0f); in test_1d() 54 VERIFY_IS_APPROX(vec5(1), 1.0f); in test_1d() 55 VERIFY_IS_APPROX(vec5(2), 2.0f * 2.0f * 2.0f); in test_1d() 56 VERIFY_IS_APPROX(vec5(3), 3.0f * 3.0f * 3.0f); in test_1d() 57 VERIFY_IS_APPROX(vec5(4), 4.0f * 4.0f * 4.0f); in test_1d() 58 VERIFY_IS_APPROX(vec5(5), 5.0f * 5.0f * 5.0f); in test_1d()
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | row_msa.cc | 478 v8i16 vec0, vec1, vec2, vec3, vec4, vec5; in I422ToRGB24Row_MSA() local 505 vec3, vec4, vec5); in I422ToRGB24Row_MSA() 508 reg3 = (v16u8)__msa_pckev_b((v16i8)vec5, (v16i8)vec2); in I422ToRGB24Row_MSA() 825 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ARGBToUVRow_MSA() local 849 vec5 = (v16u8)__msa_pckod_b((v16i8)src3, (v16i8)src2); in ARGBToUVRow_MSA() 854 vec4 = (v16u8)__msa_pckev_b((v16i8)vec5, (v16i8)vec4); in ARGBToUVRow_MSA() 855 vec5 = (v16u8)__msa_pckev_b((v16i8)vec7, (v16i8)vec6); in ARGBToUVRow_MSA() 861 reg3 = __msa_hadd_u_h(vec5, vec5); in ARGBToUVRow_MSA() 877 vec5 = (v16u8)__msa_pckod_b((v16i8)src3, (v16i8)src2); in ARGBToUVRow_MSA() 882 vec4 = (v16u8)__msa_pckev_b((v16i8)vec5, (v16i8)vec4); in ARGBToUVRow_MSA() [all …]
|
D | scale_msa.cc | 396 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_2_Box_MSA() local 421 vec5 = (v8u16)__msa_vshf_h(mask, zero, (v8i16)vec1); in ScaleRowDown38_2_Box_MSA() 428 tmp1 = __msa_hadd_u_w(vec5, vec5); in ScaleRowDown38_2_Box_MSA() 466 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_3_Box_MSA() local 489 vec5 = (v8u16)__msa_ilvl_b((v16i8)zero, (v16i8)src4); in ScaleRowDown38_3_Box_MSA() 497 vec1 += __msa_hadd_u_h((v16u8)vec5, (v16u8)vec5); in ScaleRowDown38_3_Box_MSA() 501 vec5 = (v8u16)__msa_vshf_h(mask, (v8i16)zero, (v8i16)vec1); in ScaleRowDown38_3_Box_MSA() 508 tmp1 = __msa_hadd_u_w(vec5, vec5); in ScaleRowDown38_3_Box_MSA() 565 v4i32 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ScaleFilterCols_MSA() local 586 vec5 = vec_x >> 16; in ScaleFilterCols_MSA() [all …]
|
/external/libyuv/files/source/ |
D | row_msa.cc | 478 v8i16 vec0, vec1, vec2, vec3, vec4, vec5; in I422ToRGB24Row_MSA() local 505 vec3, vec4, vec5); in I422ToRGB24Row_MSA() 508 reg3 = (v16u8)__msa_pckev_b((v16i8)vec5, (v16i8)vec2); in I422ToRGB24Row_MSA() 825 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ARGBToUVRow_MSA() local 849 vec5 = (v16u8)__msa_pckod_b((v16i8)src3, (v16i8)src2); in ARGBToUVRow_MSA() 854 vec4 = (v16u8)__msa_pckev_b((v16i8)vec5, (v16i8)vec4); in ARGBToUVRow_MSA() 855 vec5 = (v16u8)__msa_pckev_b((v16i8)vec7, (v16i8)vec6); in ARGBToUVRow_MSA() 861 reg3 = __msa_hadd_u_h(vec5, vec5); in ARGBToUVRow_MSA() 877 vec5 = (v16u8)__msa_pckod_b((v16i8)src3, (v16i8)src2); in ARGBToUVRow_MSA() 882 vec4 = (v16u8)__msa_pckev_b((v16i8)vec5, (v16i8)vec4); in ARGBToUVRow_MSA() [all …]
|
D | scale_msa.cc | 396 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_2_Box_MSA() local 421 vec5 = (v8u16)__msa_vshf_h(mask, zero, (v8i16)vec1); in ScaleRowDown38_2_Box_MSA() 428 tmp1 = __msa_hadd_u_w(vec5, vec5); in ScaleRowDown38_2_Box_MSA() 466 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_3_Box_MSA() local 489 vec5 = (v8u16)__msa_ilvl_b((v16i8)zero, (v16i8)src4); in ScaleRowDown38_3_Box_MSA() 497 vec1 += __msa_hadd_u_h((v16u8)vec5, (v16u8)vec5); in ScaleRowDown38_3_Box_MSA() 501 vec5 = (v8u16)__msa_vshf_h(mask, (v8i16)zero, (v8i16)vec1); in ScaleRowDown38_3_Box_MSA() 508 tmp1 = __msa_hadd_u_w(vec5, vec5); in ScaleRowDown38_3_Box_MSA() 565 v4i32 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ScaleFilterCols_MSA() local 586 vec5 = vec_x >> 16; in ScaleFilterCols_MSA() [all …]
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | bilinear_filter_msa.c | 56 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_4x8_msa() local 66 DOTP_UB4_UH(vec0, vec1, vec2, vec3, filt0, filt0, filt0, filt0, vec4, vec5, in common_hz_2t_4x8_msa() 68 SRARI_H4_UH(vec4, vec5, vec6, vec7, VP8_FILTER_SHIFT); in common_hz_2t_4x8_msa() 69 PCKEV_B4_SB(vec4, vec4, vec5, vec5, vec6, vec6, vec7, vec7, res0, res1, res2, in common_hz_2t_4x8_msa() 185 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_16w_msa() local 201 VSHF_B2_UB(src4, src4, src5, src5, mask, mask, vec4, vec5); in common_hz_2t_16w_msa() 205 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 225 VSHF_B2_UB(src4, src4, src5, src5, mask, mask, vec4, vec5); in common_hz_2t_16w_msa() 229 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 337 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_8x8mult_msa() local [all …]
|
/external/libaom/libaom/aom_dsp/mips/ |
D | aom_convolve8_horiz_msa.c | 347 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_4x8_msa() local 358 DOTP_UB4_UH(vec0, vec1, vec2, vec3, filt0, filt0, filt0, filt0, vec4, vec5, in common_hz_2t_4x8_msa() 360 SRARI_H4_UH(vec4, vec5, vec6, vec7, FILTER_BITS); in common_hz_2t_4x8_msa() 361 PCKEV_B4_SB(vec4, vec4, vec5, vec5, vec6, vec6, vec7, vec7, res0, res1, res2, in common_hz_2t_4x8_msa() 478 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_16w_msa() local 495 VSHF_B2_UB(src4, src4, src5, src5, mask, mask, vec4, vec5); in common_hz_2t_16w_msa() 499 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 519 VSHF_B2_UB(src4, src4, src5, src5, mask, mask, vec4, vec5); in common_hz_2t_16w_msa() 523 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 543 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_32w_msa() local [all …]
|
D | aom_convolve8_vert_msa.c | 395 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_8x8mult_msa() local 413 ILVR_B4_UB(src5, src4, src6, src5, src7, src6, src8, src7, vec4, vec5, vec6, in common_vt_2t_8x8mult_msa() 422 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, tmp0, tmp1, in common_vt_2t_8x8mult_msa() 448 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_16w_msa() local 471 ILVL_B2_UB(src3, src2, src4, src3, vec5, vec7); in common_vt_2t_16w_msa() 477 DOTP_UB2_UH(vec4, vec5, filt0, filt0, tmp0, tmp1); in common_vt_2t_16w_msa() 496 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_32w_msa() local 524 ILVL_B2_UB(src3, src2, src4, src3, vec5, vec7); in common_vt_2t_32w_msa() 525 DOTP_UB2_UH(vec4, vec5, filt0, filt0, tmp0, tmp1); in common_vt_2t_32w_msa() 544 ILVL_B2_UB(src8, src7, src9, src8, vec5, vec7); in common_vt_2t_32w_msa() [all …]
|
D | loopfilter_8_msa.c | 243 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in aom_lpf_vertical_8_dual_msa() local 280 ILVRL_H2_SH(vec1, vec0, vec4, vec5); in aom_lpf_vertical_8_dual_msa() 285 ST4x8_UB(vec4, vec5, src, pitch); in aom_lpf_vertical_8_dual_msa() 318 ILVRL_B2_SH(q2, q1, vec2, vec5); in aom_lpf_vertical_8_dual_msa() 328 ST2x4_UB(vec5, 0, src + 4, pitch); in aom_lpf_vertical_8_dual_msa() 331 ST2x4_UB(vec5, 4, src + 4, pitch); in aom_lpf_vertical_8_dual_msa()
|
/external/llvm/test/Transforms/ConstProp/ |
D | InsertElement.ll | 29 %vec5 = insertelement <4 x i64> %vec3, i64 -5, i32 4 31 ret <4 x i64> %vec5
|
/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | lowerMUL-newload.ll | 59 %vec5 = mul <4 x i32> %3, %vec3 60 %4 = trunc <4 x i32> %vec5 to <4 x i16> 68 %vec8 = add <4 x i32> %vec7, %vec5 131 %vec5 = mul <4 x i32> %3, %vec3 132 %4 = trunc <4 x i32> %vec5 to <4 x i16> 140 %vec8 = add <4 x i32> %vec7, %vec5
|
/external/llvm-project/llvm/test/Transforms/InstSimplify/ConstProp/ |
D | InsertElement.ll | 34 %vec5 = insertelement <4 x i64> %vec3, i64 -5, i32 4 35 ret <4 x i64> %vec5
|
/external/llvm-project/llvm/test/CodeGen/PowerPC/ |
D | aix-vector-stack.ll | 9 …i32> %vec1, <4 x i32> %vec2, <4 x i32> %vec3, <4 x i32> %vec4, <4 x i32> %vec5, <4 x i32> %vec6, <… 14 %add3 = add <4 x i32> %add2, %vec5
|
/external/deqp-deps/glslang/Test/ |
D | syntaxError.frag | 9 vec5 color = BaseColor;
|
/external/angle/third_party/vulkan-deps/glslang/src/Test/ |
D | syntaxError.frag | 9 vec5 color = BaseColor;
|
/external/deqp-deps/glslang/Test/baseResults/ |
D | syntaxError.frag.out | 2 ERROR: 0:9: 'vec5' : undeclared identifier
|
/external/angle/third_party/vulkan-deps/glslang/src/Test/baseResults/ |
D | syntaxError.frag.out | 2 ERROR: 0:9: 'vec5' : undeclared identifier
|