/external/libvpx/libvpx/vpx_dsp/mips/ |
D | fwd_dct32x32_msa.c | 62 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in fdct8x32_1d_column_even_store() local 72 BUTTERFLY_8(in4, in5, in6, in7, in8, in9, in10, in11, vec4, vec5, vec6, vec7, in fdct8x32_1d_column_even_store() 76 ADD4(vec0, vec7, vec1, vec6, vec2, vec5, vec3, vec4, in0, in1, in2, in3); in fdct8x32_1d_column_even_store() 88 SUB4(vec0, vec7, vec1, vec6, vec2, vec5, vec3, vec4, vec7, vec6, vec5, vec4); in fdct8x32_1d_column_even_store() 89 DOTP_CONST_PAIR(vec6, vec5, cospi_16_64, cospi_16_64, vec5, vec6); in fdct8x32_1d_column_even_store() 90 ADD2(vec4, vec5, vec7, vec6, vec0, vec1); in fdct8x32_1d_column_even_store() 96 SUB2(vec4, vec5, vec7, vec6, vec4, vec7); in fdct8x32_1d_column_even_store() 104 ADD4(in8, vec3, in9, vec2, in14, vec5, in15, vec4, in0, vec1, vec6, in2); in fdct8x32_1d_column_even_store() 105 DOTP_CONST_PAIR(vec6, vec1, cospi_24_64, cospi_8_64, in1, in3); in fdct8x32_1d_column_even_store() 291 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in fdct8x32_1d_row_even_4x() local [all …]
|
D | vpx_convolve8_avg_horiz_msa.c | 151 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_8t_and_aver_dst_16w_msa() local 173 VSHF_B4_SH(src2, src2, mask0, mask1, mask2, mask3, vec2, vec6, vec10, in common_hz_8t_and_aver_dst_16w_msa() 181 DPADD_SB4_SH(vec4, vec5, vec6, vec7, filt1, filt1, filt1, filt1, vec0, vec1, in common_hz_8t_and_aver_dst_16w_msa() 205 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_8t_and_aver_dst_32w_msa() local 229 VSHF_B4_SH(src2, src2, mask0, mask1, mask2, mask3, vec2, vec6, vec10, in common_hz_8t_and_aver_dst_32w_msa() 237 DPADD_SB4_SH(vec4, vec5, vec6, vec7, filt1, filt1, filt1, filt1, vec0, vec1, in common_hz_8t_and_aver_dst_32w_msa() 260 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_8t_and_aver_dst_64w_msa() local 286 VSHF_B4_SH(src2, src2, mask0, mask1, mask2, mask3, vec2, vec6, vec10, in common_hz_8t_and_aver_dst_64w_msa() 294 DPADD_SB4_SH(vec4, vec5, vec6, vec7, filt1, filt1, filt1, filt1, vec0, in common_hz_8t_and_aver_dst_64w_msa() 346 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_and_aver_dst_4x8_msa() local [all …]
|
D | vpx_convolve8_avg_vert_msa.c | 362 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_and_aver_dst_8x8mult_msa() local 385 ILVR_B4_UB(src5, src4, src6, src5, src7, src6, src8, src7, vec4, vec5, vec6, in common_vt_2t_and_aver_dst_8x8mult_msa() 393 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, tmp0, tmp1, in common_vt_2t_and_aver_dst_8x8mult_msa() 421 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_vt_2t_and_aver_dst_16w_msa() local 443 ILVR_B2_UB(src3, src2, src4, src3, vec4, vec6); in common_vt_2t_and_aver_dst_16w_msa() 455 DOTP_UB2_UH(vec6, vec7, filt0, filt0, tmp2, tmp3); in common_vt_2t_and_aver_dst_16w_msa() 471 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_and_aver_dst_32w_msa() local 499 ILVR_B2_UB(src3, src2, src4, src3, vec4, vec6); in common_vt_2t_and_aver_dst_32w_msa() 505 DOTP_UB2_UH(vec6, vec7, filt0, filt0, tmp2, tmp3); in common_vt_2t_and_aver_dst_32w_msa() 519 ILVR_B2_UB(src8, src7, src9, src8, vec4, vec6); in common_vt_2t_and_aver_dst_32w_msa() [all …]
|
D | vpx_convolve8_horiz_msa.c | 344 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_4x8_msa() local 356 vec6, vec7); in common_hz_2t_4x8_msa() 357 SRARI_H4_UH(vec4, vec5, vec6, vec7, FILTER_BITS); in common_hz_2t_4x8_msa() 358 PCKEV_B4_SB(vec4, vec4, vec5, vec5, vec6, vec6, vec7, vec7, res0, res1, res2, in common_hz_2t_4x8_msa() 476 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_16w_msa() local 494 VSHF_B2_UB(src6, src6, src7, src7, mask, mask, vec6, vec7); in common_hz_2t_16w_msa() 497 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 518 VSHF_B2_UB(src6, src6, src7, src7, mask, mask, vec6, vec7); in common_hz_2t_16w_msa() 521 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 541 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_32w_msa() local [all …]
|
D | vpx_convolve8_vert_msa.c | 394 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_8x8mult_msa() local 412 ILVR_B4_UB(src5, src4, src6, src5, src7, src6, src8, src7, vec4, vec5, vec6, in common_vt_2t_8x8mult_msa() 421 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, tmp0, tmp1, in common_vt_2t_8x8mult_msa() 447 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_16w_msa() local 469 ILVR_B2_UB(src3, src2, src4, src3, vec4, vec6); in common_vt_2t_16w_msa() 481 DOTP_UB2_UH(vec6, vec7, filt0, filt0, tmp2, tmp3); in common_vt_2t_16w_msa() 495 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_32w_msa() local 522 ILVR_B2_UB(src3, src2, src4, src3, vec4, vec6); in common_vt_2t_32w_msa() 528 DOTP_UB2_UH(vec6, vec7, filt0, filt0, tmp2, tmp3); in common_vt_2t_32w_msa() 542 ILVR_B2_UB(src8, src7, src9, src8, vec4, vec6); in common_vt_2t_32w_msa() [all …]
|
D | loopfilter_8_msa.c | 243 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in vpx_lpf_vertical_8_dual_msa() local 317 ILVRL_H2_SH(vec1, vec0, vec6, vec7); in vpx_lpf_vertical_8_dual_msa() 327 ST4x4_UB(vec6, vec6, 0, 1, 2, 3, src, pitch); in vpx_lpf_vertical_8_dual_msa()
|
D | sub_pixel_variance_msa.c | 472 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in sub_pixel_sse_diff_16width_h_msa() local 490 VSHF_B2_UH(src6, src6, src7, src7, mask, mask, vec6, vec7); in sub_pixel_sse_diff_16width_h_msa() 493 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in sub_pixel_sse_diff_16width_h_msa() 642 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in sub_pixel_sse_diff_16width_v_msa() local 666 ILVR_B2_UB(src3, src2, src4, src3, vec4, vec6); in sub_pixel_sse_diff_16width_v_msa() 675 DOTP_UB2_UH(vec6, vec7, filt0, filt0, tmp2, tmp3); in sub_pixel_sse_diff_16width_v_msa() 1054 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in subpel_avg_ssediff_16w_h_msa() local 1074 VSHF_B2_UH(src6, src6, src7, src7, mask, mask, vec6, vec7); in subpel_avg_ssediff_16w_h_msa() 1077 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in subpel_avg_ssediff_16w_h_msa() 1253 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in subpel_avg_ssediff_16w_v_msa() local [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | row_msa.cc | 825 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ARGBToUVRow_MSA() local 850 vec6 = (v16u8)__msa_pckod_b((v16i8)src5, (v16i8)src4); in ARGBToUVRow_MSA() 855 vec5 = (v16u8)__msa_pckev_b((v16i8)vec7, (v16i8)vec6); in ARGBToUVRow_MSA() 878 vec6 = (v16u8)__msa_pckod_b((v16i8)src5, (v16i8)src4); in ARGBToUVRow_MSA() 883 vec5 = (v16u8)__msa_pckev_b((v16i8)vec7, (v16i8)vec6); in ARGBToUVRow_MSA() 986 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ARGBToRGB565Row_MSA() local 997 vec6 = (v16u8)__msa_srai_b((v16i8)src1, 5); in ARGBToRGB565Row_MSA() 1001 vec6 = (v16u8)__msa_sldi_b(zero, (v16i8)vec6, 1); in ARGBToRGB565Row_MSA() 1007 vec5 = __msa_binsli_b(vec6, vec7, 4); in ARGBToRGB565Row_MSA() 1022 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ARGBToARGB1555Row_MSA() local [all …]
|
D | scale_msa.cc | 396 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_2_Box_MSA() local 422 vec6 = (v8u16)__msa_vshf_h(mask, zero, (v8i16)vec2); in ScaleRowDown38_2_Box_MSA() 429 tmp2 = __msa_hadd_u_w(vec6, vec6); in ScaleRowDown38_2_Box_MSA() 466 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_3_Box_MSA() local 490 vec6 = (v8u16)__msa_ilvr_b((v16i8)zero, (v16i8)src5); in ScaleRowDown38_3_Box_MSA() 498 vec2 += __msa_hadd_u_h((v16u8)vec6, (v16u8)vec6); in ScaleRowDown38_3_Box_MSA() 502 vec6 = (v8u16)__msa_vshf_h(mask, (v8i16)zero, (v8i16)vec2); in ScaleRowDown38_3_Box_MSA() 509 tmp2 = __msa_hadd_u_w(vec6, vec6); in ScaleRowDown38_3_Box_MSA() 565 v4i32 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ScaleFilterCols_MSA() local 578 vec6 = vec_x & const_0xFFFF; in ScaleFilterCols_MSA() [all …]
|
/external/libyuv/files/source/ |
D | row_msa.cc | 825 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ARGBToUVRow_MSA() local 850 vec6 = (v16u8)__msa_pckod_b((v16i8)src5, (v16i8)src4); in ARGBToUVRow_MSA() 855 vec5 = (v16u8)__msa_pckev_b((v16i8)vec7, (v16i8)vec6); in ARGBToUVRow_MSA() 878 vec6 = (v16u8)__msa_pckod_b((v16i8)src5, (v16i8)src4); in ARGBToUVRow_MSA() 883 vec5 = (v16u8)__msa_pckev_b((v16i8)vec7, (v16i8)vec6); in ARGBToUVRow_MSA() 986 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ARGBToRGB565Row_MSA() local 997 vec6 = (v16u8)__msa_srai_b((v16i8)src1, 5); in ARGBToRGB565Row_MSA() 1001 vec6 = (v16u8)__msa_sldi_b(zero, (v16i8)vec6, 1); in ARGBToRGB565Row_MSA() 1007 vec5 = __msa_binsli_b(vec6, vec7, 4); in ARGBToRGB565Row_MSA() 1022 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ARGBToARGB1555Row_MSA() local [all …]
|
D | scale_msa.cc | 396 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_2_Box_MSA() local 422 vec6 = (v8u16)__msa_vshf_h(mask, zero, (v8i16)vec2); in ScaleRowDown38_2_Box_MSA() 429 tmp2 = __msa_hadd_u_w(vec6, vec6); in ScaleRowDown38_2_Box_MSA() 466 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in ScaleRowDown38_3_Box_MSA() local 490 vec6 = (v8u16)__msa_ilvr_b((v16i8)zero, (v16i8)src5); in ScaleRowDown38_3_Box_MSA() 498 vec2 += __msa_hadd_u_h((v16u8)vec6, (v16u8)vec6); in ScaleRowDown38_3_Box_MSA() 502 vec6 = (v8u16)__msa_vshf_h(mask, (v8i16)zero, (v8i16)vec2); in ScaleRowDown38_3_Box_MSA() 509 tmp2 = __msa_hadd_u_w(vec6, vec6); in ScaleRowDown38_3_Box_MSA() 565 v4i32 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, vec8, vec9; in ScaleFilterCols_MSA() local 578 vec6 = vec_x & const_0xFFFF; in ScaleFilterCols_MSA() [all …]
|
/external/libvpx/libvpx/vp8/common/mips/msa/ |
D | bilinear_filter_msa.c | 56 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_4x8_msa() local 67 vec6, vec7); in common_hz_2t_4x8_msa() 68 SRARI_H4_UH(vec4, vec5, vec6, vec7, VP8_FILTER_SHIFT); in common_hz_2t_4x8_msa() 69 PCKEV_B4_SB(vec4, vec4, vec5, vec5, vec6, vec6, vec7, vec7, res0, res1, res2, in common_hz_2t_4x8_msa() 185 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_16w_msa() local 202 VSHF_B2_UB(src6, src6, src7, src7, mask, mask, vec6, vec7); in common_hz_2t_16w_msa() 205 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 226 VSHF_B2_UB(src6, src6, src7, src7, mask, mask, vec6, vec7); in common_hz_2t_16w_msa() 229 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 337 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_8x8mult_msa() local [all …]
|
/external/libaom/libaom/aom_dsp/mips/ |
D | aom_convolve8_horiz_msa.c | 347 v8u16 vec4, vec5, vec6, vec7, filt; in common_hz_2t_4x8_msa() local 359 vec6, vec7); in common_hz_2t_4x8_msa() 360 SRARI_H4_UH(vec4, vec5, vec6, vec7, FILTER_BITS); in common_hz_2t_4x8_msa() 361 PCKEV_B4_SB(vec4, vec4, vec5, vec5, vec6, vec6, vec7, vec7, res0, res1, res2, in common_hz_2t_4x8_msa() 478 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_16w_msa() local 496 VSHF_B2_UB(src6, src6, src7, src7, mask, mask, vec6, vec7); in common_hz_2t_16w_msa() 499 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 520 VSHF_B2_UB(src6, src6, src7, src7, mask, mask, vec6, vec7); in common_hz_2t_16w_msa() 523 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in common_hz_2t_16w_msa() 543 v16u8 filt0, vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in common_hz_2t_32w_msa() local [all …]
|
D | aom_convolve8_vert_msa.c | 395 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_8x8mult_msa() local 413 ILVR_B4_UB(src5, src4, src6, src5, src7, src6, src8, src7, vec4, vec5, vec6, in common_vt_2t_8x8mult_msa() 422 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, tmp0, tmp1, in common_vt_2t_8x8mult_msa() 448 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_16w_msa() local 470 ILVR_B2_UB(src3, src2, src4, src3, vec4, vec6); in common_vt_2t_16w_msa() 482 DOTP_UB2_UH(vec6, vec7, filt0, filt0, tmp2, tmp3); in common_vt_2t_16w_msa() 496 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7, filt0; in common_vt_2t_32w_msa() local 523 ILVR_B2_UB(src3, src2, src4, src3, vec4, vec6); in common_vt_2t_32w_msa() 529 DOTP_UB2_UH(vec6, vec7, filt0, filt0, tmp2, tmp3); in common_vt_2t_32w_msa() 543 ILVR_B2_UB(src8, src7, src9, src8, vec4, vec6); in common_vt_2t_32w_msa() [all …]
|
D | loopfilter_8_msa.c | 243 v8i16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in aom_lpf_vertical_8_dual_msa() local 317 ILVRL_H2_SH(vec1, vec0, vec6, vec7); in aom_lpf_vertical_8_dual_msa() 327 ST4x4_UB(vec6, vec6, 0, 1, 2, 3, src, pitch); in aom_lpf_vertical_8_dual_msa()
|
D | sub_pixel_variance_msa.c | 469 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in sub_pixel_sse_diff_16width_h_msa() local 487 VSHF_B2_UH(src6, src6, src7, src7, mask, mask, vec6, vec7); in sub_pixel_sse_diff_16width_h_msa() 490 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in sub_pixel_sse_diff_16width_h_msa() 639 v16u8 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in sub_pixel_sse_diff_16width_v_msa() local 663 ILVR_B2_UB(src3, src2, src4, src3, vec4, vec6); in sub_pixel_sse_diff_16width_v_msa() 672 DOTP_UB2_UH(vec6, vec7, filt0, filt0, tmp2, tmp3); in sub_pixel_sse_diff_16width_v_msa() 1051 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in subpel_avg_ssediff_16w_h_msa() local 1071 VSHF_B2_UH(src6, src6, src7, src7, mask, mask, vec6, vec7); in subpel_avg_ssediff_16w_h_msa() 1074 DOTP_UB4_UH(vec4, vec5, vec6, vec7, filt0, filt0, filt0, filt0, out4, out5, in subpel_avg_ssediff_16w_h_msa() 1250 v8u16 vec0, vec1, vec2, vec3, vec4, vec5, vec6, vec7; in subpel_avg_ssediff_16w_v_msa() local [all …]
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | vec_loadsingles.ll | 53 %vec6 = insertelement <4 x float> %vec4, float %d, i32 3 54 ret <4 x float> %vec6 83 %vec6 = insertelement <8 x float> %vec4, float %d, i32 3 86 %vec8 = insertelement <8 x float> %vec6, float %e, i32 4 120 %vec6 = insertelement <4 x double> %vec4, double %d, i32 3 121 ret <4 x double> %vec6
|
/external/llvm/test/CodeGen/X86/ |
D | vec_loadsingles.ll | 53 %vec6 = insertelement <4 x float> %vec4, float %d, i32 3 54 ret <4 x float> %vec6 83 %vec6 = insertelement <8 x float> %vec4, float %d, i32 3 86 %vec8 = insertelement <8 x float> %vec6, float %e, i32 4 120 %vec6 = insertelement <4 x double> %vec4, double %d, i32 3 121 ret <4 x double> %vec6
|
/external/llvm-project/llvm/test/CodeGen/PowerPC/ |
D | aix-vector-stack.ll | 9 …i32> %vec2, <4 x i32> %vec3, <4 x i32> %vec4, <4 x i32> %vec5, <4 x i32> %vec6, <4 x i32> %vec7, <… 15 %add4 = add <4 x i32> %add3, %vec6
|
D | aix-cc-ext-vec-abi.ll | 9 …i32> %vec2, <4 x i32> %vec3, <4 x i32> %vec4, <4 x i32> %vec5, <4 x i32> %vec6, <4 x i32> %vec7, <… 15 %add4 = add <4 x i32> %add3, %vec6
|
/external/llvm-project/llvm/test/CodeGen/ARM/ |
D | legalize-bitcast.ll | 32 %vec6 = load <6 x i16>, <6 x i16>* @vec6_p, align 1 36 %0 = bitcast <6 x i16> %vec6 to i96
|
D | lowerMUL-newload.ll | 65 %vec6 = load <4 x i16>, <4 x i16>* %vector_ptr6, align 8 66 %6 = sext <4 x i16> %vec6 to <4 x i32> 137 %vec6 = load <4 x i16>, <4 x i16>* %vector_ptr6, align 8 138 %6 = sext <4 x i16> %vec6 to <4 x i32>
|
/external/llvm-project/polly/docs/experiments/matmul/ |
D | matmul.polly.interchanged+tiled+vector+openmp.ll | 274 …%strided.vec6 = shufflevector <16 x float> %wide.vec15, <16 x float> undef, <4 x i32> <i32 2, i32 … 288 %21 = fadd <4 x float> %strided.vec6, %20 296 …%strided.vec6.1 = shufflevector <16 x float> %wide.vec.117, <16 x float> undef, <4 x i32> <i32 2, … 310 %33 = fadd <4 x float> %strided.vec6.1, %32 318 …%strided.vec6.2 = shufflevector <16 x float> %wide.vec.219, <16 x float> undef, <4 x i32> <i32 2, … 332 %45 = fadd <4 x float> %strided.vec6.2, %44 340 …%strided.vec6.3 = shufflevector <16 x float> %wide.vec.321, <16 x float> undef, <4 x i32> <i32 2, … 354 %57 = fadd <4 x float> %strided.vec6.3, %56
|
/external/arm-trusted-firmware/plat/mediatek/mt8173/drivers/spm/ |
D | spm.h | 269 unsigned int vec6; /* event vector 6 config */ member
|
D | spm.c | 231 mmio_write_32(SPM_PCM_EVENT_VECTOR6, pcmdesc->vec6); in spm_init_event_vector()
|