/third_party/ffmpeg/libavcodec/ |
D | me_cmp.c | 146 #define avg2(a, b) (((a) + (b) + 1) >> 1) macro 222 s += abs(pix1[0] - avg2(pix2[0], pix2[1])); in pix_abs16_x2_c() 223 s += abs(pix1[1] - avg2(pix2[1], pix2[2])); in pix_abs16_x2_c() 224 s += abs(pix1[2] - avg2(pix2[2], pix2[3])); in pix_abs16_x2_c() 225 s += abs(pix1[3] - avg2(pix2[3], pix2[4])); in pix_abs16_x2_c() 226 s += abs(pix1[4] - avg2(pix2[4], pix2[5])); in pix_abs16_x2_c() 227 s += abs(pix1[5] - avg2(pix2[5], pix2[6])); in pix_abs16_x2_c() 228 s += abs(pix1[6] - avg2(pix2[6], pix2[7])); in pix_abs16_x2_c() 229 s += abs(pix1[7] - avg2(pix2[7], pix2[8])); in pix_abs16_x2_c() 230 s += abs(pix1[8] - avg2(pix2[8], pix2[9])); in pix_abs16_x2_c() [all …]
|
D | vc1dsp.c | 833 #define avg2(a, b) (((a) + (b) + 1) >> 1) macro 847 dst[0] = avg2(dst[0], chroma_mc(0)); in avg_no_rnd_vc1_chroma_mc8_c() 848 dst[1] = avg2(dst[1], chroma_mc(1)); in avg_no_rnd_vc1_chroma_mc8_c() 849 dst[2] = avg2(dst[2], chroma_mc(2)); in avg_no_rnd_vc1_chroma_mc8_c() 850 dst[3] = avg2(dst[3], chroma_mc(3)); in avg_no_rnd_vc1_chroma_mc8_c() 851 dst[4] = avg2(dst[4], chroma_mc(4)); in avg_no_rnd_vc1_chroma_mc8_c() 852 dst[5] = avg2(dst[5], chroma_mc(5)); in avg_no_rnd_vc1_chroma_mc8_c() 853 dst[6] = avg2(dst[6], chroma_mc(6)); in avg_no_rnd_vc1_chroma_mc8_c() 854 dst[7] = avg2(dst[7], chroma_mc(7)); in avg_no_rnd_vc1_chroma_mc8_c() 873 dst[0] = avg2(dst[0], chroma_mc(0)); in avg_no_rnd_vc1_chroma_mc4_c() [all …]
|
/third_party/ffmpeg/libavcodec/alpha/ |
D | me_cmp_alpha.c | 28 static inline uint64_t avg2(uint64_t a, uint64_t b) in avg2() function 137 p2_l = avg2(l, (l >> 8) | ((uint64_t) r << 56)); in pix_abs16x16_x2_mvi() 138 p2_r = avg2(r, (r >> 8) | ((uint64_t) pix2[16] << 56)); in pix_abs16x16_x2_mvi() 160 p2_l = avg2(extql(l, disalign) | extqh(m, disalign), m); in pix_abs16x16_x2_mvi() 161 p2_r = avg2(extql(m, disalign) | extqh(r, disalign), r); in pix_abs16x16_x2_mvi() 180 p2_l = avg2(extql(l, disalign) | extqh(m, disalign), in pix_abs16x16_x2_mvi() 182 p2_r = avg2(extql(m, disalign) | extqh(r, disalign), in pix_abs16x16_x2_mvi() 216 result += perr(p1_l, avg2(p2_l, np2_l)) in pix_abs16x16_y2_mvi() 217 + perr(p1_r, avg2(p2_r, np2_r)); in pix_abs16x16_y2_mvi() 237 result += perr(p1_l, avg2(p2_l, np2_l)) in pix_abs16x16_y2_mvi() [all …]
|
D | hpeldsp_alpha.c | 32 static inline uint64_t avg2(uint64_t a, uint64_t b) in avg2() function 145 #define AVG2 avg2
|
/third_party/gstreamer/gstplugins_bad/gst/debugutils/ |
D | gstcompare.c | 360 gdouble avg1, avg2, var1, var2, cov; in gst_compare_ssim_window() local 391 avg2 = sum2 / count; in gst_compare_ssim_window() 393 var2 = ssum2 / count - avg2 * avg2; in gst_compare_ssim_window() 394 cov = acov / count - avg1 * avg2; in gst_compare_ssim_window() 396 return (2 * avg1 * avg2 + c1) * (2 * cov + c2) / in gst_compare_ssim_window() 397 ((avg1 * avg1 + avg2 * avg2 + c1) * (var1 + var2 + c2)); in gst_compare_ssim_window()
|
/third_party/flutter/skia/third_party/externals/libwebp/src/dsp/ |
D | dec_sse2.c | 926 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in LD4_SSE2() local 927 const __m128i abcdefg = _mm_avg_epu8(avg2, BCDEFGH0); in LD4_SSE2() 947 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in VR4_SSE2() local 948 const __m128i efgh = _mm_avg_epu8(avg2, XABCD); in VR4_SSE2() 965 const __m128i avg2 = _mm_avg_epu8(CDEFGH__, BCDEFGH_); in VL4_SSE2() local 966 const __m128i avg3 = _mm_avg_epu8(avg1, avg2); in VL4_SSE2() 967 const __m128i lsb1 = _mm_and_si128(_mm_xor_si128(avg1, avg2), one); in VL4_SSE2() 999 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in RD4_SSE2() local 1000 const __m128i abcdefg = _mm_avg_epu8(avg2, KJIXABCD_); in RD4_SSE2()
|
D | lossless_sse2.c | 125 const __m128i avg2 = _mm_srli_epi16(sum, 1); in Average3_SSE2() local 126 const __m128i A2 = _mm_packus_epi16(avg2, avg2); in Average3_SSE2() 134 const __m128i avg2 = Average2_uint32_16_SSE2(a2, a3); in Average4_SSE2() local 135 const __m128i sum = _mm_add_epi16(avg2, avg1); in Average4_SSE2()
|
D | enc_sse2.c | 761 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in LD4_SSE2() local 762 const __m128i abcdefg = _mm_avg_epu8(avg2, BCDEFGH0); in LD4_SSE2() 783 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in VR4_SSE2() local 784 const __m128i efgh = _mm_avg_epu8(avg2, XABCD); in VR4_SSE2() 802 const __m128i avg2 = _mm_avg_epu8(CDEFGH__, BCDEFGH_); in VL4_SSE2() local 803 const __m128i avg3 = _mm_avg_epu8(avg1, avg2); in VL4_SSE2() 804 const __m128i lsb1 = _mm_and_si128(_mm_xor_si128(avg1, avg2), one); in VL4_SSE2() 830 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in RD4_SSE2() local 831 const __m128i abcdefg = _mm_avg_epu8(avg2, KJIXABCD_); in RD4_SSE2()
|
D | dec_neon.c | 1372 const uint8x8_t avg2 = vrhadd_u8(avg1, KJIXABC_); in RD4_NEON() local 1373 const uint64x1_t avg2_u64 = vreinterpret_u64_u8(avg2); in RD4_NEON() 1374 const uint32x2_t r3 = vreinterpret_u32_u8(avg2); in RD4_NEON() 1391 const uint8x8_t avg2 = vrhadd_u8(avg1, BCDEFGH0); in LD4_NEON() local 1392 const uint64x1_t avg2_u64 = vreinterpret_u64_u8(avg2); in LD4_NEON() 1393 const uint32x2_t r0 = vreinterpret_u32_u8(avg2); in LD4_NEON()
|
/third_party/skia/third_party/externals/libwebp/src/dsp/ |
D | dec_sse2.c | 926 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in LD4_SSE2() local 927 const __m128i abcdefg = _mm_avg_epu8(avg2, BCDEFGH0); in LD4_SSE2() 947 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in VR4_SSE2() local 948 const __m128i efgh = _mm_avg_epu8(avg2, XABCD); in VR4_SSE2() 965 const __m128i avg2 = _mm_avg_epu8(CDEFGH__, BCDEFGH_); in VL4_SSE2() local 966 const __m128i avg3 = _mm_avg_epu8(avg1, avg2); in VL4_SSE2() 967 const __m128i lsb1 = _mm_and_si128(_mm_xor_si128(avg1, avg2), one); in VL4_SSE2() 999 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in RD4_SSE2() local 1000 const __m128i abcdefg = _mm_avg_epu8(avg2, KJIXABCD_); in RD4_SSE2()
|
D | lossless_sse2.c | 124 const __m128i avg2 = _mm_srli_epi16(sum, 1); in Average3_SSE2() local 125 const __m128i A2 = _mm_packus_epi16(avg2, avg2); in Average3_SSE2() 133 const __m128i avg2 = Average2_uint32_16_SSE2(a2, a3); in Average4_SSE2() local 134 const __m128i sum = _mm_add_epi16(avg2, avg1); in Average4_SSE2()
|
D | enc_sse2.c | 761 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in LD4_SSE2() local 762 const __m128i abcdefg = _mm_avg_epu8(avg2, BCDEFGH0); in LD4_SSE2() 783 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in VR4_SSE2() local 784 const __m128i efgh = _mm_avg_epu8(avg2, XABCD); in VR4_SSE2() 802 const __m128i avg2 = _mm_avg_epu8(CDEFGH__, BCDEFGH_); in VL4_SSE2() local 803 const __m128i avg3 = _mm_avg_epu8(avg1, avg2); in VL4_SSE2() 804 const __m128i lsb1 = _mm_and_si128(_mm_xor_si128(avg1, avg2), one); in VL4_SSE2() 830 const __m128i avg2 = _mm_subs_epu8(avg1, lsb); in RD4_SSE2() local 831 const __m128i abcdefg = _mm_avg_epu8(avg2, KJIXABCD_); in RD4_SSE2()
|
D | dec_neon.c | 1373 const uint8x8_t avg2 = vrhadd_u8(avg1, KJIXABC_); in RD4_NEON() local 1374 const uint64x1_t avg2_u64 = vreinterpret_u64_u8(avg2); in RD4_NEON() 1375 const uint32x2_t r3 = vreinterpret_u32_u8(avg2); in RD4_NEON() 1392 const uint8x8_t avg2 = vrhadd_u8(avg1, BCDEFGH0); in LD4_NEON() local 1393 const uint64x1_t avg2_u64 = vreinterpret_u64_u8(avg2); in LD4_NEON() 1394 const uint32x2_t r0 = vreinterpret_u32_u8(avg2); in LD4_NEON()
|