/external/libvpx/libvpx/vpx_dsp/arm/ |
D | variance_neon.c | 213 uint32_t sse1, sse2; in vpx_variance32x64_neon() local 214 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 32, 32, &sse1, in vpx_variance32x64_neon() 219 *sse = sse1 + sse2; in vpx_variance32x64_neon() 228 uint32_t sse1, sse2; in vpx_variance64x32_neon() local 229 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x32_neon() 234 *sse = sse1 + sse2; in vpx_variance64x32_neon() 243 uint32_t sse1, sse2; in vpx_variance64x64_neon() local 245 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x64_neon() 250 sse1 += sse2; in vpx_variance64x64_neon() 256 sse1 += sse2; in vpx_variance64x64_neon() [all …]
|
/external/libaom/libaom/aom_dsp/arm/ |
D | variance_neon.c | 102 uint32_t sse1, sse2; in aom_variance32x64_neon() local 103 variance_neon_w8(a, a_stride, b, b_stride, 32, 32, &sse1, &sum1); in aom_variance32x64_neon() 106 *sse = sse1 + sse2; in aom_variance32x64_neon() 115 uint32_t sse1, sse2; in aom_variance64x32_neon() local 116 variance_neon_w8(a, a_stride, b, b_stride, 64, 16, &sse1, &sum1); in aom_variance64x32_neon() 119 *sse = sse1 + sse2; in aom_variance64x32_neon() 128 uint32_t sse1, sse2; in aom_variance64x64_neon() local 130 variance_neon_w8(a, a_stride, b, b_stride, 64, 16, &sse1, &sum1); in aom_variance64x64_neon() 133 sse1 += sse2; in aom_variance64x64_neon() 138 sse1 += sse2; in aom_variance64x64_neon() [all …]
|
/external/libaom/libaom/test/ |
D | variance_test.cc | 516 unsigned int sse1, sse2, var1, var2; in RefTest() local 519 var1 = params_.func(src_, stride, ref_, stride, &sse1)); in RefTest() 523 EXPECT_EQ(sse1, sse2) << "Error at test index: " << i; in RefTest() 544 unsigned int sse1, sse2; in RefStrideTest() local 548 var1 = params_.func(src_, src_stride, ref_, ref_stride, &sse1)); in RefStrideTest() 552 EXPECT_EQ(sse1, sse2) << "Error at test index: " << i; in RefStrideTest() 612 unsigned int sse1, sse2; in RefTestMse() local 614 ASM_REGISTER_STATE_CHECK(params_.func(src_, stride, ref_, stride, &sse1)); in RefTestMse() 617 EXPECT_EQ(sse1, sse2); in RefTestMse() 743 unsigned int sse1, sse2; in RefTest() local [all …]
|
/external/python/cpython2/Modules/_ctypes/libffi_osx/x86/ |
D | x86-ffi64.c | 381 _Bool sse1 = n == 2 && SSE_CLASS_P(classes[1]); in ffi_prep_cif_machdep() local 383 if (sse0 && !sse1) in ffi_prep_cif_machdep() 385 else if (!sse0 && sse1) in ffi_prep_cif_machdep() 387 else if (sse0 && sse1) in ffi_prep_cif_machdep() 654 _Bool sse1 = SSE_CLASS_P (classes[1]); in ffi_closure_unix64_inner() local 656 if (!sse0 && sse1) in ffi_closure_unix64_inner() 658 else if (sse0 && !sse1) in ffi_closure_unix64_inner()
|
/external/python/cpython2/Modules/_ctypes/libffi/src/x86/ |
D | ffi64.c | 379 _Bool sse1 = n == 2 && SSE_CLASS_P (classes[1]); in ffi_prep_cif_machdep() local 380 if (sse0 && !sse1) in ffi_prep_cif_machdep() 382 else if (!sse0 && sse1) in ffi_prep_cif_machdep() 384 else if (sse0 && sse1) in ffi_prep_cif_machdep() 602 _Bool sse1 = SSE_CLASS_P (classes[1]); in ffi_closure_unix64_inner() local 603 if (!sse0 && sse1) in ffi_closure_unix64_inner() 605 else if (sse0 && !sse1) in ffi_closure_unix64_inner()
|
/external/python/cpython3/Modules/_ctypes/libffi_osx/x86/ |
D | x86-ffi64.c | 383 _Bool sse1 = n == 2 && SSE_CLASS_P(classes[1]); in ffi_prep_cif_machdep() local 385 if (sse0 && !sse1) in ffi_prep_cif_machdep() 387 else if (!sse0 && sse1) in ffi_prep_cif_machdep() 389 else if (sse0 && sse1) in ffi_prep_cif_machdep() 657 _Bool sse1 = SSE_CLASS_P (classes[1]); in ffi_closure_unix64_inner() local 659 if (!sse0 && sse1) in ffi_closure_unix64_inner() 661 else if (sse0 && !sse1) in ffi_closure_unix64_inner()
|
/external/libffi/src/x86/ |
D | ffi64.c | 379 _Bool sse1 = n == 2 && SSE_CLASS_P (classes[1]); in ffi_prep_cif_machdep() local 380 if (sse0 && !sse1) in ffi_prep_cif_machdep() 382 else if (!sse0 && sse1) in ffi_prep_cif_machdep() 384 else if (sse0 && sse1) in ffi_prep_cif_machdep() 602 _Bool sse1 = SSE_CLASS_P (classes[1]); in ffi_closure_unix64_inner() local 603 if (!sse0 && sse1) in ffi_closure_unix64_inner() 605 else if (sse0 && !sse1) in ffi_closure_unix64_inner()
|
/external/libvpx/libvpx/test/ |
D | variance_test.cc | 402 unsigned int sse1, sse2, var1, var2; in RefTest() local 405 var1 = params_.func(src_, stride, ref_, stride, &sse1)); in RefTest() 409 EXPECT_EQ(sse1, sse2) << "Error at test index: " << i; in RefTest() 432 unsigned int sse1, sse2; in RefStrideTest() local 436 var1 = params_.func(src_, src_stride, ref_, ref_stride, &sse1)); in RefStrideTest() 440 EXPECT_EQ(sse1, sse2) << "Error at test index: " << i; in RefStrideTest() 505 unsigned int sse1, sse2; in RefTestMse() local 507 ASM_REGISTER_STATE_CHECK(params_.func(src_, stride, ref_, stride, &sse1)); in RefTestMse() 510 EXPECT_EQ(sse1, sse2); in RefTestMse() 636 unsigned int sse1, sse2; in RefTest() local [all …]
|
/external/libvpx/libvpx/vp8/encoder/ |
D | mcomp.c | 216 *sse1 = sse; \ 226 unsigned int *sse1) { in vp8_find_best_sub_pixel_step_iteratively() argument 290 besterr = vfp->vf(y, y_stride, z, b->src_stride, sse1); in vp8_find_best_sub_pixel_step_iteratively() 370 unsigned int *sse1) { in vp8_find_best_sub_pixel_step() argument 405 bestmse = vfp->vf(y, y_stride, z, b->src_stride, sse1); in vp8_find_best_sub_pixel_step() 420 *sse1 = sse; in vp8_find_best_sub_pixel_step() 432 *sse1 = sse; in vp8_find_best_sub_pixel_step() 446 *sse1 = sse; in vp8_find_best_sub_pixel_step() 458 *sse1 = sse; in vp8_find_best_sub_pixel_step() 500 *sse1 = sse; in vp8_find_best_sub_pixel_step() [all …]
|
D | rdopt.c | 377 unsigned int sse1 = 0; in VP8_UVSSE() local 407 vpred_ptr, uv_stride, &sse1); in VP8_UVSSE() 408 sse2 += sse1; in VP8_UVSSE() 411 vpx_variance8x8(vptr, pre_stride, vpred_ptr, uv_stride, &sse1); in VP8_UVSSE() 412 sse2 += sse1; in VP8_UVSSE()
|
/external/libaom/libaom/av1/encoder/ |
D | mcomp.c | 205 *sse1 = sse; \ 228 *sse1 = sse; \ 338 int *mvcost[2], unsigned int *sse1, int *distortion) { in setup_center_error() argument 351 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); in setup_center_error() 360 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); in setup_center_error() 363 besterr = vfp->vf(y + offset, y_stride, src, src_stride, sse1); in setup_center_error() 399 unsigned int *sse1, const uint8_t *second_pred, const uint8_t *mask, in av1_find_best_sub_pixel_tree_pruned_evenmore() argument 406 h, offset, mvjcost, mvcost, sse1, distortion); in av1_find_best_sub_pixel_tree_pruned_evenmore() 472 unsigned int *sse1, const uint8_t *second_pred, const uint8_t *mask, in av1_find_best_sub_pixel_tree_pruned_more() argument 485 h, offset, mvjcost, mvcost, sse1, distortion); in av1_find_best_sub_pixel_tree_pruned_more() [all …]
|
D | mcomp.h | 110 int *distortion, unsigned int *sse1, const uint8_t *second_pred, 154 int *mvjcost, int *mvcost[2], int *distortion, unsigned int *sse1,
|
/external/libvpx/libvpx/vp9/encoder/ |
D | vp9_mcomp.c | 188 *sse1 = sse; \ 211 *sse1 = sse; \ 305 int *mvjcost, int *mvcost[2], uint32_t *sse1, uint32_t *distortion) { in setup_center_error() argument 314 vfp->vf(CONVERT_TO_BYTEPTR(comp_pred16), w, src, src_stride, sse1); in setup_center_error() 318 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); in setup_center_error() 321 besterr = vfp->vf(y + offset, y_stride, src, src_stride, sse1); in setup_center_error() 333 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); in setup_center_error() 335 besterr = vfp->vf(y + offset, y_stride, src, src_stride, sse1); in setup_center_error() 374 uint32_t *distortion, uint32_t *sse1, const uint8_t *second_pred, int w, in vp9_skip_sub_pixel_tree() argument 379 offset, mvjcost, mvcost, sse1, distortion); in vp9_skip_sub_pixel_tree() [all …]
|
D | vp9_mcomp.h | 77 uint32_t *distortion, uint32_t *sse1, const uint8_t *second_pred, int w,
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | variance_avx2.c | 741 unsigned int sse1; in vpx_sub_pixel_variance64x64_avx2() local 743 src_ptr, src_stride, x_offset, y_offset, ref_ptr, ref_stride, 64, &sse1); in vpx_sub_pixel_variance64x64_avx2() 749 *sse = sse1 + sse2; in vpx_sub_pixel_variance64x64_avx2() 765 unsigned int sse1; in vpx_sub_pixel_avg_variance64x64_avx2() local 768 second_pred, 64, 64, &sse1); in vpx_sub_pixel_avg_variance64x64_avx2() 775 *sse = sse1 + sse2; in vpx_sub_pixel_avg_variance64x64_avx2()
|