/third_party/openh264/codec/common/loongarch/ |
D | mc_horver_lsx.c | 124 __m128i src4, src5, src6, src7; in PixelAvgWidthEq16_lsx() local 137 src5 = __lsx_vld(pSrcB, 0); in PixelAvgWidthEq16_lsx() 140 src4 = __lsx_vavgr_bu(src4, src5); in PixelAvgWidthEq16_lsx() 163 __m128i src0, src1, src2, src3, src4, src5; in McHorVer02WidthEq8_lsx() local 172 src5 = __lsx_vldx(psrc, iStride3); in McHorVer02WidthEq8_lsx() 180 src5 = __lsx_vsllwil_hu_bu(src5, 0); in McHorVer02WidthEq8_lsx() 181 FILTER_INPUT_8BIT(src0, src1, src2, src3 ,src4, src5 ,src0); in McHorVer02WidthEq8_lsx() 197 __m128i src0, src1, src2, src3, src4, src5; in McHorVer02WidthEq16_lsx() local 207 src5 = __lsx_vldx(psrc, iStride3); in McHorVer02WidthEq16_lsx() 216 tmp5 = __lsx_vsllwil_hu_bu(src5, 0); in McHorVer02WidthEq16_lsx() [all …]
|
D | copy_mb_lsx.c | 51 __m128i src0, src1, src2, src3, src4 ,src5, src6, src7; in WelsCopy8x8_lsx() local 65 src4, src5, src6, src7); in WelsCopy8x8_lsx() 76 __lsx_vstelm_d(src5, pDst + iStrideD, 0, 0); in WelsCopy8x8_lsx() 90 __m128i src0, src1, src2, src3, src4, src5, src6, src7; in WelsCopy16x16_lsx() local 101 src4, src5, src6, src7); in WelsCopy16x16_lsx() 124 __lsx_vstx(src5, pDst, iStride1); in WelsCopy16x16_lsx() 147 v16u8_b src0, src1, src2, src3, src4, src5, src6, src7; in WelsCopy16x16NotAligned_lsx() local 158 src4, src5, src6, src7); in WelsCopy16x16NotAligned_lsx() 181 __lsx_vstx((__m128i)src5, pDst, iStride1); in WelsCopy16x16NotAligned_lsx()
|
/third_party/openh264/codec/decoder/core/loongarch/ |
D | mb_aux_lsx.c | 196 __m128i src0, src1, src2, src3, src4, src5, src6, src7; in IdctResAddPred8x8_lsx() local 211 src4, src5, src6, src7); in IdctResAddPred8x8_lsx() 225 LSX_TRANSPOSE8x8_H(src0, src1, src2, src3, src4, src5, src6, src7, in IdctResAddPred8x8_lsx() 226 src0, src1, src2, src3, src4, src5, src6, src7); in IdctResAddPred8x8_lsx() 227 DO(src0, src1, src2, src3, src4, src5, src6, src7, in IdctResAddPred8x8_lsx() 228 src0, src1, src2, src3, src4, src5, src6, src7); in IdctResAddPred8x8_lsx() 230 LSX_TRANSPOSE8x8_H(src0, src1, src2, src3, src4, src5, src6, src7, in IdctResAddPred8x8_lsx() 231 src0, src1, src2, src3, src4, src5, src6, src7); in IdctResAddPred8x8_lsx() 232 DO(src0, src1, src2, src3, src4, src5, src6, src7, in IdctResAddPred8x8_lsx() 233 src0, src1, src2, src3, src4, src5, src6, src7); in IdctResAddPred8x8_lsx() [all …]
|
/third_party/ffmpeg/libavcodec/mips/ |
D | h264qpel_msa.c | 647 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in ff_put_h264_qpel16_mc00_msa() local 650 LD_UB8(src, stride, src0, src1, src2, src3, src4, src5, src6, src7); in ff_put_h264_qpel16_mc00_msa() 654 ST_UB8(src0, src1, src2, src3, src4, src5, src6, src7, dst, stride); in ff_put_h264_qpel16_mc00_msa() 662 uint64_t src0, src1, src2, src3, src4, src5, src6, src7; in ff_put_h264_qpel8_mc00_msa() local 666 LD4(src, stride, src4, src5, src6, src7); in ff_put_h264_qpel8_mc00_msa() 669 SD4(src4, src5, src6, src7, dst, stride); in ff_put_h264_qpel8_mc00_msa() 675 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in ff_avg_h264_qpel16_mc00_msa() local 678 LD_UB8(src, stride, src0, src1, src2, src3, src4, src5, src6, src7); in ff_avg_h264_qpel16_mc00_msa() 684 AVER_UB4_UB(src4, dst4, src5, dst5, src6, dst6, src7, dst7, dst4, dst5, in ff_avg_h264_qpel16_mc00_msa() 689 LD_UB8(src, stride, src0, src1, src2, src3, src4, src5, src6, src7); in ff_avg_h264_qpel16_mc00_msa() [all …]
|
D | hpeldsp_msa.c | 105 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in common_hz_bil_16w_msa() local 109 LD_UB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in common_hz_bil_16w_msa() 118 AVER_ST16x4_UB(src4, src12, src5, src13, src6, src14, src7, src15, in common_hz_bil_16w_msa() 127 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in common_hz_bil_no_rnd_8x8_msa() local 132 LD_SB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in common_hz_bil_no_rnd_8x8_msa() 137 SLDI_B4_SB(zeros, src4, zeros, src5, zeros, src6, zeros, src7, 1, in common_hz_bil_no_rnd_8x8_msa() 143 AVE_ST8x4_UB(src4, src4_sld1, src5, src5_sld1, in common_hz_bil_no_rnd_8x8_msa() 164 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; in common_hz_bil_no_rnd_16x16_msa() local 167 LD_UB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in common_hz_bil_no_rnd_16x16_msa() 180 AVE_ST16x4_UB(src4, src12, src5, src13, src6, src14, src7, src15, in common_hz_bil_no_rnd_16x16_msa() [all …]
|
D | hevc_mc_uni_msa.c | 150 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in copy_width12_msa() local 152 LD_UB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in copy_width12_msa() 154 ST12x8_UB(src0, src1, src2, src3, src4, src5, src6, src7, dst, dst_stride); in copy_width12_msa() 156 LD_UB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in copy_width12_msa() 157 ST12x8_UB(src0, src1, src2, src3, src4, src5, src6, src7, dst, dst_stride); in copy_width12_msa() 165 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in copy_width16_msa() local 168 LD_UB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in copy_width16_msa() 170 ST_UB8(src0, src1, src2, src3, src4, src5, src6, src7, dst, dst_stride); in copy_width16_msa() 178 LD_UB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, in copy_width16_msa() 181 ST_UB8(src0, src1, src2, src3, src4, src5, src6, src7, dst, in copy_width16_msa() [all …]
|
D | hevcdsp_msa.c | 60 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_copy_4w_msa() local 66 src0, src1, src2, src3, src4, src5, src6, src7); in hevc_copy_4w_msa() 69 ILVR_W4_SB(src1, src0, src3, src2, src5, src4, src7, src6, in hevc_copy_4w_msa() 86 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_copy_6w_msa() local 90 LD_SB8(src, src_stride, src0, src1, src2, src3, src4, src5, src6, src7); in hevc_copy_6w_msa() 95 ILVR_B4_SH(zero, src4, zero, src5, zero, src6, zero, src7, in hevc_copy_6w_msa() 131 v16i8 src0, src1, src2, src3, src4, src5; in hevc_copy_8w_msa() local 134 LD_SB6(src, src_stride, src0, src1, src2, src3, src4, src5); in hevc_copy_8w_msa() 138 ILVR_B2_SH(zero, src4, zero, src5, in4, in5); in hevc_copy_8w_msa() 145 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_copy_8w_msa() local [all …]
|
D | vp8_mc_msa.c | 289 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, filt0, filt1, filt2; in ff_put_vp8_epel16_h6_msa() local 305 LD_SB4(src + 8, src_stride, src1, src3, src5, src7); in ff_put_vp8_epel16_h6_msa() 306 XORI_B8_128_SB(src0, src1, src2, src3, src4, src5, src6, src7); in ff_put_vp8_epel16_h6_msa() 311 HORIZ_6TAP_8WID_4VECS_FILT(src4, src5, src6, src7, mask0, mask1, mask2, in ff_put_vp8_epel16_h6_msa() 338 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, src8; in ff_put_vp8_epel4_v6_msa() local 358 LD_SB4(src, src_stride, src5, src6, src7, src8); in ff_put_vp8_epel4_v6_msa() 361 ILVR_B4_SB(src5, src4, src6, src5, src7, src6, src8, src7, src54_r, in ff_put_vp8_epel4_v6_msa() 435 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, src8; in ff_put_vp8_epel16_v6_msa() local 457 LD_SB4(src, src_stride, src5, src6, src7, src8); in ff_put_vp8_epel16_v6_msa() 460 XORI_B4_128_SB(src5, src6, src7, src8); in ff_put_vp8_epel16_v6_msa() [all …]
|
D | vp9_mc_msa.c | 475 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, src10; in common_vt_8t_4w_msa() local 487 LD_SB7(src, src_stride, src0, src1, src2, src3, src4, src5, src6); in common_vt_8t_4w_msa() 490 ILVR_B4_SB(src1, src0, src3, src2, src5, src4, src2, src1, src10_r, src32_r, in common_vt_8t_4w_msa() 492 ILVR_B2_SB(src4, src3, src6, src5, src43_r, src65_r); in common_vt_8t_4w_msa() 527 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, src10; in common_vt_8t_8w_msa() local 538 LD_SB7(src, src_stride, src0, src1, src2, src3, src4, src5, src6); in common_vt_8t_8w_msa() 539 XORI_B7_128_SB(src0, src1, src2, src3, src4, src5, src6); in common_vt_8t_8w_msa() 541 ILVR_B4_SB(src1, src0, src3, src2, src5, src4, src2, src1, src10_r, src32_r, in common_vt_8t_8w_msa() 543 ILVR_B2_SB(src4, src3, src6, src5, src43_r, src65_r); in common_vt_8t_8w_msa() 582 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, src10; in common_vt_8t_16w_msa() local [all …]
|
D | hevc_mc_uniw_msa.c | 380 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_uniwgt_copy_24w_msa() local 392 LD_SB4(src, src_stride, src0, src1, src4, src5); in hevc_uniwgt_copy_24w_msa() 400 ILVRL_B2_SH(zero, src5, dst8, dst9); in hevc_uniwgt_copy_24w_msa() 479 v16i8 src0, src1, src2, src3, src4, src5; in hevc_uniwgt_copy_48w_msa() local 493 LD_SB3(src, 16, src3, src4, src5); in hevc_uniwgt_copy_48w_msa() 501 ILVRL_B2_SH(zero, src5, dst10, dst11); in hevc_uniwgt_copy_48w_msa() 536 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_uniwgt_copy_64w_msa() local 550 LD_SB4(src, 16, src4, src5, src6, src7); in hevc_uniwgt_copy_64w_msa() 558 ILVRL_B2_SH(zero, src5, dst10, dst11); in hevc_uniwgt_copy_64w_msa() 601 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_hz_uniwgt_8t_4w_msa() local [all …]
|
D | hevc_mc_bi_msa.c | 356 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, zero = { 0 }; in hevc_bi_copy_24w_msa() local 361 LD_SB4(src0_ptr, src_stride, src0, src1, src4, src5); in hevc_bi_copy_24w_msa() 373 ILVRL_B2_SH(zero, src5, dst8, dst9); in hevc_bi_copy_24w_msa() 446 v16i8 src0, src1, src2, src3, src4, src5; in hevc_bi_copy_48w_msa() local 454 LD_SB3(src0_ptr, 16, src3, src4, src5); in hevc_bi_copy_48w_msa() 467 ILVRL_B2_SH(zero, src5, dst10, dst11); in hevc_bi_copy_48w_msa() 540 v16i8 src0, src1, src2, src3, src4, src5, src6, src7; in hevc_hz_bi_8t_4w_msa() local 563 src4, src5, src6, src7); in hevc_hz_bi_8t_4w_msa() 570 XORI_B8_128_SB(src0, src1, src2, src3, src4, src5, src6, src7); in hevc_hz_bi_8t_4w_msa() 577 VSHF_B2_SB(src4, src5, src6, src7, mask0, mask0, vec2, vec3); in hevc_hz_bi_8t_4w_msa() [all …]
|
D | pixblockdsp_msa.c | 72 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in copy_16multx8mult_msa() local 80 src0, src1, src2, src3, src4, src5, src6, src7); in copy_16multx8mult_msa() 83 ST_UB8(src0, src1, src2, src3, src4, src5, src6, src7, in copy_16multx8mult_msa() 98 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in copy_width16_msa() local 103 src0, src1, src2, src3, src4, src5, src6, src7); in copy_width16_msa() 105 ST_UB8(src0, src1, src2, src3, src4, src5, src6, src7, in copy_width16_msa()
|
D | h264chroma_msa.c | 153 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, out0, out1; in avc_chroma_hz_4x8_msa() local 162 LD_UB8(src, stride, src0, src1, src2, src3, src4, src5, src6, src7); in avc_chroma_hz_4x8_msa() 164 VSHF_B2_UB(src4, src5, src6, src7, mask, mask, src4, src6); in avc_chroma_hz_4x8_msa() 213 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in avc_chroma_hz_8x8_msa() local 223 LD_UB8(src, stride, src0, src1, src2, src3, src4, src5, src6, src7); in avc_chroma_hz_8x8_msa() 226 VSHF_B2_UB(src4, src4, src5, src5, mask, mask, src4, src5); in avc_chroma_hz_8x8_msa() 230 DOTP_UB4_UH(src4, src5, src6, src7, coeff_vec, coeff_vec, coeff_vec, in avc_chroma_hz_8x8_msa() 427 v16u8 src0, src1, src2, src3, src4, src5, src6, src7, src8; in avc_chroma_vt_4x8_msa() local 436 LD_UB4(src, stride, src5, src6, src7, src8); in avc_chroma_vt_4x8_msa() 439 ILVR_B4_UB(src5, src4, src6, src5, src7, src6, src8, src7, tmp4, tmp5, tmp6, in avc_chroma_vt_4x8_msa() [all …]
|
D | hevc_mc_biw_msa.c | 443 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, zero = { 0 }; in hevc_biwgt_copy_24w_msa() local 457 LD_SB4(src0_ptr, src_stride, src0, src1, src4, src5); in hevc_biwgt_copy_24w_msa() 469 ILVRL_B2_SH(zero, src5, dst8, dst9); in hevc_biwgt_copy_24w_msa() 1417 v16i8 src0, src1, src2, src3, src4, src5, src6, src7, src8, src9, src10; in hevc_vt_biwgt_8t_4w_msa() local 1446 LD_SB7(src0_ptr, src_stride, src0, src1, src2, src3, src4, src5, src6); in hevc_vt_biwgt_8t_4w_msa() 1449 ILVR_B4_SB(src1, src0, src3, src2, src5, src4, src2, src1, in hevc_vt_biwgt_8t_4w_msa() 1451 ILVR_B2_SB(src4, src3, src6, src5, src43_r, src65_r); in hevc_vt_biwgt_8t_4w_msa() 1515 v16i8 src0, src1, src2, src3, src4, src5; in hevc_vt_biwgt_8t_8w_msa() local 1541 LD_SB7(src0_ptr, src_stride, src0, src1, src2, src3, src4, src5, src6); in hevc_vt_biwgt_8t_8w_msa() 1543 XORI_B7_128_SB(src0, src1, src2, src3, src4, src5, src6); in hevc_vt_biwgt_8t_8w_msa() [all …]
|
D | ac3dsp_mips.c | 208 float src0, src1, src2, src3, src4, src5, src6, src7; in float_to_fixed24_mips() local 257 [src4] "=&f" (src4), [src5] "=&f" (src5), in float_to_fixed24_mips()
|
/third_party/skia/third_party/externals/libpng/mips/ |
D | filter_msa_intrinsics.c | 374 v16u8 src0, src1, src2, src3, src4, src5, src6, src7; in png_read_filter_row_up_msa() local 379 LD_UB4(pp, 16, src4, src5, src6, src7); in png_read_filter_row_up_msa() 382 ADD4(src0, src4, src1, src5, src2, src6, src3, src7, in png_read_filter_row_up_msa() 400 LD_UB4(pp, 16, src4, src5, src6, src7); in png_read_filter_row_up_msa() 402 ADD4(src0, src4, src1, src5, src2, src6, src3, src7, in png_read_filter_row_up_msa() 411 LD_UB2(pp, 16, src4, src5); in png_read_filter_row_up_msa() 416 ADD3(src0, src4, src1, src5, src2, src6, src0, src1, src2); in png_read_filter_row_up_msa() 426 LD_UB2(pp, 16, src4, src5); in png_read_filter_row_up_msa() 428 ADD2(src0, src4, src1, src5, src0, src1); in png_read_filter_row_up_msa() 437 LD_UB2(pp, 16, src4, src5); in png_read_filter_row_up_msa() [all …]
|
/third_party/ffmpeg/libavcodec/ppc/ |
D | vc1dsp_altivec.c | 140 vector signed short src0, src1, src2, src3, src4, src5, src6, src7; in vc1_inv_trans_8x8_altivec() local 158 src5 = vec_ld( 80, block); in vc1_inv_trans_8x8_altivec() 167 s5 = vec_unpackl(src5); in vc1_inv_trans_8x8_altivec() 175 sD = vec_unpackh(src5); in vc1_inv_trans_8x8_altivec() 187 src5 = vec_pack(sD, s5); in vc1_inv_trans_8x8_altivec() 190 TRANSPOSE8(src0, src1, src2, src3, src4, src5, src6, src7); in vc1_inv_trans_8x8_altivec() 197 s5 = vec_unpackl(src5); in vc1_inv_trans_8x8_altivec() 205 sD = vec_unpackh(src5); in vc1_inv_trans_8x8_altivec() 217 src5 = vec_pack(sD, s5); in vc1_inv_trans_8x8_altivec() 226 vec_st(src5, 80, block); in vc1_inv_trans_8x8_altivec() [all …]
|
/third_party/openh264/codec/encoder/core/loongarch/ |
D | dct_lasx.c | 71 __m256i src0, src1, src2, src3, src4, src5, src6, src7; in WelsDctT4_lasx() local 85 src4, src5, src6, src7); in WelsDctT4_lasx() 88 src1, src5, in WelsDctT4_lasx() 125 __m256i src0, src1, src2, src3, src4, src5, src6, src7, in WelsDctFourT4_lasx() local 141 src4, src5, src6, src7); in WelsDctFourT4_lasx() 156 src1, src5, in WelsDctFourT4_lasx() 237 __m256i src0, src1, src2, src3, src4, src5, src6, src7, in WelsIDctT4Rec_lasx() local 244 src5 = __lasx_xvldx(pPred, iPredStride); in WelsIDctT4Rec_lasx() 299 src4, src5, src6, src7, in WelsIDctT4Rec_lasx()
|
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Hexagon/ |
D | HexagonMapAsm2IntrinV65.gen.td | 70 …egs:$src3, HvxVR:$src4, HvxVR:$src5), (V6_vscattermwq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, H… 71 …egs:$src3, HvxVR:$src4, HvxVR:$src5), (V6_vscattermhq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, H… 74 …egs:$src3, HvxWR:$src4, HvxVR:$src5), (V6_vscattermhwq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, … 79 …egs:$src3, HvxVR:$src4, HvxVR:$src5), (V6_vscattermwq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, H… 80 …egs:$src3, HvxVR:$src4, HvxVR:$src5), (V6_vscattermhq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, H… 83 …egs:$src3, HvxWR:$src4, HvxVR:$src5), (V6_vscattermhwq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, …
|
D | HexagonIntrinsics.td | 444 …t<(int_hexagon_V6_vscattermwq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxVR:$src4, HvxVR:$src5), 445 …(V6_vscattermwq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxVR:$src4, HvxVR:$src5)>, Requires<[H… 446 …t<(int_hexagon_V6_vscattermhq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxVR:$src4, HvxVR:$src5), 447 …(V6_vscattermhq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxVR:$src4, HvxVR:$src5)>, Requires<[H… 452 …<(int_hexagon_V6_vscattermhwq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxWR:$src4, HvxVR:$src5), 453 …(V6_vscattermhwq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxWR:$src4, HvxVR:$src5)>, Requires<[… 462 …t_hexagon_V6_vscattermwq_128B HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxVR:$src4, HvxVR:$src5), 463 …(V6_vscattermwq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxVR:$src4, HvxVR:$src5)>, Requires<[H… 464 …t_hexagon_V6_vscattermhq_128B HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxVR:$src4, HvxVR:$src5), 465 …(V6_vscattermhq HvxQR:$src1, IntRegs:$src2, ModRegs:$src3, HvxVR:$src4, HvxVR:$src5)>, Requires<[H… [all …]
|
/third_party/ffmpeg/libavcodec/ |
D | qpeldsp.c | 77 const int src5 = src[5 * srcStride]; \ 82 …(dst[1 * dstStride], (src1 + src2) * 20 - (src0 + src3) * 6 + (src0 + src4) * 3 - (src1 + src5)); \ 83 …OP(dst[2 * dstStride], (src2 + src3) * 20 - (src1 + src4) * 6 + (src0 + src5) * 3 - (src0 + src6))… 84 …OP(dst[3 * dstStride], (src3 + src4) * 20 - (src2 + src5) * 6 + (src1 + src6) * 3 - (src0 + src7))… 85 …OP(dst[4 * dstStride], (src4 + src5) * 20 - (src3 + src6) * 6 + (src2 + src7) * 3 - (src1 + src8))… 86 …OP(dst[5 * dstStride], (src5 + src6) * 20 - (src4 + src7) * 6 + (src3 + src8) * 3 - (src2 + src8))… 87 …OP(dst[6 * dstStride], (src6 + src7) * 20 - (src5 + src8) * 6 + (src4 + src8) * 3 - (src3 + src7))… 88 …OP(dst[7 * dstStride], (src7 + src8) * 20 - (src6 + src8) * 6 + (src5 + src7) * 3 - (src4 + src6))… 138 const int src5 = src[5 * srcStride]; \ 151 … dstStride], (src1 + src2) * 20 - (src0 + src3) * 6 + (src0 + src4) * 3 - (src1 + src5)); \ [all …]
|
D | wmv2dsp.c | 175 const int src5 = src[5 * srcStride]; in wmv2_mspel8_v_lowpass() local 183 dst[3 * dstStride] = cm[(9 * (src3 + src4) - (src2 + src5) + 8) >> 4]; in wmv2_mspel8_v_lowpass() 184 dst[4 * dstStride] = cm[(9 * (src4 + src5) - (src3 + src6) + 8) >> 4]; in wmv2_mspel8_v_lowpass() 185 dst[5 * dstStride] = cm[(9 * (src5 + src6) - (src4 + src7) + 8) >> 4]; in wmv2_mspel8_v_lowpass() 186 dst[6 * dstStride] = cm[(9 * (src6 + src7) - (src5 + src8) + 8) >> 4]; in wmv2_mspel8_v_lowpass()
|
D | h264qpel_template.c | 189 const int src5= src[5 *srcStride];\ 193 OP(dst[2*dstStride], (src2+src3)*20 - (src1+src4)*5 + (src0+src5));\ 194 OP(dst[3*dstStride], (src3+src4)*20 - (src2+src5)*5 + (src1+src6));\ 281 const int src5= src[5 *srcStride];\ 289 OP(dst[2*dstStride], (src2+src3)*20 - (src1+src4)*5 + (src0+src5));\ 290 OP(dst[3*dstStride], (src3+src4)*20 - (src2+src5)*5 + (src1+src6));\ 291 OP(dst[4*dstStride], (src4+src5)*20 - (src3+src6)*5 + (src2+src7));\ 292 OP(dst[5*dstStride], (src5+src6)*20 - (src4+src7)*5 + (src3+src8));\ 293 OP(dst[6*dstStride], (src6+src7)*20 - (src5+src8)*5 + (src4+src9));\ 294 OP(dst[7*dstStride], (src7+src8)*20 - (src6+src9)*5 + (src5+src10));\
|
/third_party/ffmpeg/libavcodec/x86/ |
D | cavsidct.asm | 36 mova m2, [%1+5*16] ; m2 = src5 49 paddw m6, m6 ; m6 = 2*src5 53 paddw m2, m6 ; m2 = 3*src5 56 paddw m7, m6 ; m7 = 3*src3 - 2*src5 = a1 57 psubw m1, m2 ; m1 = 2*src3 - 3*src5 = a2
|
/third_party/ffmpeg/libavcodec/aarch64/ |
D | vp9mc_neon.S | 165 .macro extmla dst1, dst2, dst3, dst4, src1, src2, src3, src4, src5, src6, offset, size 167 ext v22.16b, \src4\().16b, \src5\().16b, #(2*\offset) 172 ext v23.16b, \src5\().16b, \src6\().16b, #(2*\offset) 185 .macro extmulqadd dst1, dst2, dst3, dst4, src1, src2, src3, src4, src5, src6, offset, size 187 ext v22.16b, \src4\().16b, \src5\().16b, #(2*\offset) 192 ext v23.16b, \src5\().16b, \src6\().16b, #(2*\offset) 433 .macro convolve dst1, dst2, src1, src2, src3, src4, src5, src6, src7, src8, src9, idx1, idx2, tmp1,… 442 mla \dst2\().8h, \src5\().8h, v0.h[3] 444 mla \dst1\().8h, \src5\().8h, v0.h[4] 455 mla \tmp2\().8h, \src5\().8h, v0.h[3] [all …]
|