Searched refs:vtmp3 (Results 1 – 5 of 5) sorted by relevance
/third_party/skia/third_party/externals/libpng/arm/ |
D | filter_neon_intrinsics.c | 172 uint8x8_t vtmp1, vtmp2, vtmp3; in png_read_filter_row_avg3_neon() local 188 vtmp3 = vext_u8(vrp.val[0], vrp.val[1], 6); in png_read_filter_row_avg3_neon() 200 vdest.val[2] = vadd_u8(vdest.val[2], vtmp3); in png_read_filter_row_avg3_neon() 312 uint8x8_t vtmp1, vtmp2, vtmp3; in png_read_filter_row_paeth3_neon() local 328 vtmp3 = vext_u8(vpp.val[0], vpp.val[1], 6); in png_read_filter_row_paeth3_neon() 329 vdest.val[2] = paeth(vdest.val[1], vtmp3, vtmp2); in png_read_filter_row_paeth3_neon() 339 vdest.val[3] = paeth(vdest.val[2], vtmp2, vtmp3); in png_read_filter_row_paeth3_neon()
|
/third_party/optimized-routines/networking/aarch64/ |
D | chksum_simd.c | 89 uint32x4_t vtmp3 = vld1q_u32(ptr32 + 12); in __chksum_aarch64_simd() local 93 vsum3 = vpadalq_u32(vsum3, vtmp3); in __chksum_aarch64_simd()
|
/third_party/optimized-routines/networking/arm/ |
D | chksum_simd.c | 62 uint32x4_t vtmp3 = vld1q_u32(ptr32 + 12); in __chksum_arm_simd() local 66 vsum3 = vpadalq_u32(vsum3, vtmp3); in __chksum_arm_simd()
|
/third_party/ffmpeg/libavcodec/ppc/ |
D | h264dsp.c | 90 vec_s16 vtmp0, vtmp1, vtmp2, vtmp3; in h264_idct_add_altivec() local 105 vtmp3 = vec_sld(vtmp2, vtmp2, 8); in h264_idct_add_altivec() 108 VEC_1D_DCT(vtmp0,vtmp1,vtmp2,vtmp3,va0,va1,va2,va3); in h264_idct_add_altivec() 109 VEC_TRANSPOSE_4(va0,va1,va2,va3,vtmp0,vtmp1,vtmp2,vtmp3); in h264_idct_add_altivec() 110 VEC_1D_DCT(vtmp0,vtmp1,vtmp2,vtmp3,va0,va1,va2,va3); in h264_idct_add_altivec()
|
/third_party/ffmpeg/libswscale/ppc/ |
D | swscale_vsx.c | 1676 vec_s16 vtmp, vtmp2, vtmp3, vtmp4; in hyscale_fast_vsx() local 1741 vtmp3 = vec_mergeh(vtmp, vtmp2); in hyscale_fast_vsx() 1749 vd_l = vec_add(vd_l, (vec_u16) vtmp3); in hyscale_fast_vsx() 1770 vtmp3 = vec_mergeh(vtmp, vtmp2); \ 1778 vd_l = vec_add(vd_l, vtmp3); \ 1792 vec_u16 vtmp, vtmp2, vtmp3, vtmp4; in hcscale_fast_vsx() local
|