Home
last modified time | relevance | path

Searched refs:sse1 (Results 1 – 15 of 15) sorted by relevance

/external/libvpx/libvpx/vpx_dsp/arm/
Dvariance_neon.c213 uint32_t sse1, sse2; in vpx_variance32x64_neon() local
214 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 32, 32, &sse1, in vpx_variance32x64_neon()
219 *sse = sse1 + sse2; in vpx_variance32x64_neon()
228 uint32_t sse1, sse2; in vpx_variance64x32_neon() local
229 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x32_neon()
234 *sse = sse1 + sse2; in vpx_variance64x32_neon()
243 uint32_t sse1, sse2; in vpx_variance64x64_neon() local
245 variance_neon_w16(src_ptr, src_stride, ref_ptr, ref_stride, 64, 16, &sse1, in vpx_variance64x64_neon()
250 sse1 += sse2; in vpx_variance64x64_neon()
256 sse1 += sse2; in vpx_variance64x64_neon()
[all …]
/external/libaom/libaom/aom_dsp/arm/
Dvariance_neon.c102 uint32_t sse1, sse2; in aom_variance32x64_neon() local
103 variance_neon_w8(a, a_stride, b, b_stride, 32, 32, &sse1, &sum1); in aom_variance32x64_neon()
106 *sse = sse1 + sse2; in aom_variance32x64_neon()
115 uint32_t sse1, sse2; in aom_variance64x32_neon() local
116 variance_neon_w8(a, a_stride, b, b_stride, 64, 16, &sse1, &sum1); in aom_variance64x32_neon()
119 *sse = sse1 + sse2; in aom_variance64x32_neon()
128 uint32_t sse1, sse2; in aom_variance64x64_neon() local
130 variance_neon_w8(a, a_stride, b, b_stride, 64, 16, &sse1, &sum1); in aom_variance64x64_neon()
133 sse1 += sse2; in aom_variance64x64_neon()
138 sse1 += sse2; in aom_variance64x64_neon()
[all …]
/external/libaom/libaom/test/
Dvariance_test.cc516 unsigned int sse1, sse2, var1, var2; in RefTest() local
519 var1 = params_.func(src_, stride, ref_, stride, &sse1)); in RefTest()
523 EXPECT_EQ(sse1, sse2) << "Error at test index: " << i; in RefTest()
544 unsigned int sse1, sse2; in RefStrideTest() local
548 var1 = params_.func(src_, src_stride, ref_, ref_stride, &sse1)); in RefStrideTest()
552 EXPECT_EQ(sse1, sse2) << "Error at test index: " << i; in RefStrideTest()
612 unsigned int sse1, sse2; in RefTestMse() local
614 ASM_REGISTER_STATE_CHECK(params_.func(src_, stride, ref_, stride, &sse1)); in RefTestMse()
617 EXPECT_EQ(sse1, sse2); in RefTestMse()
743 unsigned int sse1, sse2; in RefTest() local
[all …]
/external/python/cpython2/Modules/_ctypes/libffi_osx/x86/
Dx86-ffi64.c381 _Bool sse1 = n == 2 && SSE_CLASS_P(classes[1]); in ffi_prep_cif_machdep() local
383 if (sse0 && !sse1) in ffi_prep_cif_machdep()
385 else if (!sse0 && sse1) in ffi_prep_cif_machdep()
387 else if (sse0 && sse1) in ffi_prep_cif_machdep()
654 _Bool sse1 = SSE_CLASS_P (classes[1]); in ffi_closure_unix64_inner() local
656 if (!sse0 && sse1) in ffi_closure_unix64_inner()
658 else if (sse0 && !sse1) in ffi_closure_unix64_inner()
/external/python/cpython2/Modules/_ctypes/libffi/src/x86/
Dffi64.c379 _Bool sse1 = n == 2 && SSE_CLASS_P (classes[1]); in ffi_prep_cif_machdep() local
380 if (sse0 && !sse1) in ffi_prep_cif_machdep()
382 else if (!sse0 && sse1) in ffi_prep_cif_machdep()
384 else if (sse0 && sse1) in ffi_prep_cif_machdep()
602 _Bool sse1 = SSE_CLASS_P (classes[1]); in ffi_closure_unix64_inner() local
603 if (!sse0 && sse1) in ffi_closure_unix64_inner()
605 else if (sse0 && !sse1) in ffi_closure_unix64_inner()
/external/python/cpython3/Modules/_ctypes/libffi_osx/x86/
Dx86-ffi64.c383 _Bool sse1 = n == 2 && SSE_CLASS_P(classes[1]); in ffi_prep_cif_machdep() local
385 if (sse0 && !sse1) in ffi_prep_cif_machdep()
387 else if (!sse0 && sse1) in ffi_prep_cif_machdep()
389 else if (sse0 && sse1) in ffi_prep_cif_machdep()
657 _Bool sse1 = SSE_CLASS_P (classes[1]); in ffi_closure_unix64_inner() local
659 if (!sse0 && sse1) in ffi_closure_unix64_inner()
661 else if (sse0 && !sse1) in ffi_closure_unix64_inner()
/external/libffi/src/x86/
Dffi64.c379 _Bool sse1 = n == 2 && SSE_CLASS_P (classes[1]); in ffi_prep_cif_machdep() local
380 if (sse0 && !sse1) in ffi_prep_cif_machdep()
382 else if (!sse0 && sse1) in ffi_prep_cif_machdep()
384 else if (sse0 && sse1) in ffi_prep_cif_machdep()
602 _Bool sse1 = SSE_CLASS_P (classes[1]); in ffi_closure_unix64_inner() local
603 if (!sse0 && sse1) in ffi_closure_unix64_inner()
605 else if (sse0 && !sse1) in ffi_closure_unix64_inner()
/external/libvpx/libvpx/test/
Dvariance_test.cc402 unsigned int sse1, sse2, var1, var2; in RefTest() local
405 var1 = params_.func(src_, stride, ref_, stride, &sse1)); in RefTest()
409 EXPECT_EQ(sse1, sse2) << "Error at test index: " << i; in RefTest()
432 unsigned int sse1, sse2; in RefStrideTest() local
436 var1 = params_.func(src_, src_stride, ref_, ref_stride, &sse1)); in RefStrideTest()
440 EXPECT_EQ(sse1, sse2) << "Error at test index: " << i; in RefStrideTest()
505 unsigned int sse1, sse2; in RefTestMse() local
507 ASM_REGISTER_STATE_CHECK(params_.func(src_, stride, ref_, stride, &sse1)); in RefTestMse()
510 EXPECT_EQ(sse1, sse2); in RefTestMse()
636 unsigned int sse1, sse2; in RefTest() local
[all …]
/external/libvpx/libvpx/vp8/encoder/
Dmcomp.c216 *sse1 = sse; \
226 unsigned int *sse1) { in vp8_find_best_sub_pixel_step_iteratively() argument
290 besterr = vfp->vf(y, y_stride, z, b->src_stride, sse1); in vp8_find_best_sub_pixel_step_iteratively()
370 unsigned int *sse1) { in vp8_find_best_sub_pixel_step() argument
405 bestmse = vfp->vf(y, y_stride, z, b->src_stride, sse1); in vp8_find_best_sub_pixel_step()
420 *sse1 = sse; in vp8_find_best_sub_pixel_step()
432 *sse1 = sse; in vp8_find_best_sub_pixel_step()
446 *sse1 = sse; in vp8_find_best_sub_pixel_step()
458 *sse1 = sse; in vp8_find_best_sub_pixel_step()
500 *sse1 = sse; in vp8_find_best_sub_pixel_step()
[all …]
Drdopt.c377 unsigned int sse1 = 0; in VP8_UVSSE() local
407 vpred_ptr, uv_stride, &sse1); in VP8_UVSSE()
408 sse2 += sse1; in VP8_UVSSE()
411 vpx_variance8x8(vptr, pre_stride, vpred_ptr, uv_stride, &sse1); in VP8_UVSSE()
412 sse2 += sse1; in VP8_UVSSE()
/external/libaom/libaom/av1/encoder/
Dmcomp.c205 *sse1 = sse; \
228 *sse1 = sse; \
338 int *mvcost[2], unsigned int *sse1, int *distortion) { in setup_center_error() argument
351 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); in setup_center_error()
360 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); in setup_center_error()
363 besterr = vfp->vf(y + offset, y_stride, src, src_stride, sse1); in setup_center_error()
399 unsigned int *sse1, const uint8_t *second_pred, const uint8_t *mask, in av1_find_best_sub_pixel_tree_pruned_evenmore() argument
406 h, offset, mvjcost, mvcost, sse1, distortion); in av1_find_best_sub_pixel_tree_pruned_evenmore()
472 unsigned int *sse1, const uint8_t *second_pred, const uint8_t *mask, in av1_find_best_sub_pixel_tree_pruned_more() argument
485 h, offset, mvjcost, mvcost, sse1, distortion); in av1_find_best_sub_pixel_tree_pruned_more()
[all …]
Dmcomp.h110 int *distortion, unsigned int *sse1, const uint8_t *second_pred,
154 int *mvjcost, int *mvcost[2], int *distortion, unsigned int *sse1,
/external/libvpx/libvpx/vp9/encoder/
Dvp9_mcomp.c188 *sse1 = sse; \
211 *sse1 = sse; \
305 int *mvjcost, int *mvcost[2], uint32_t *sse1, uint32_t *distortion) { in setup_center_error() argument
314 vfp->vf(CONVERT_TO_BYTEPTR(comp_pred16), w, src, src_stride, sse1); in setup_center_error()
318 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); in setup_center_error()
321 besterr = vfp->vf(y + offset, y_stride, src, src_stride, sse1); in setup_center_error()
333 besterr = vfp->vf(comp_pred, w, src, src_stride, sse1); in setup_center_error()
335 besterr = vfp->vf(y + offset, y_stride, src, src_stride, sse1); in setup_center_error()
374 uint32_t *distortion, uint32_t *sse1, const uint8_t *second_pred, int w, in vp9_skip_sub_pixel_tree() argument
379 offset, mvjcost, mvcost, sse1, distortion); in vp9_skip_sub_pixel_tree()
[all …]
Dvp9_mcomp.h77 uint32_t *distortion, uint32_t *sse1, const uint8_t *second_pred, int w,
/external/libvpx/libvpx/vpx_dsp/x86/
Dvariance_avx2.c741 unsigned int sse1; in vpx_sub_pixel_variance64x64_avx2() local
743 src_ptr, src_stride, x_offset, y_offset, ref_ptr, ref_stride, 64, &sse1); in vpx_sub_pixel_variance64x64_avx2()
749 *sse = sse1 + sse2; in vpx_sub_pixel_variance64x64_avx2()
765 unsigned int sse1; in vpx_sub_pixel_avg_variance64x64_avx2() local
768 second_pred, 64, 64, &sse1); in vpx_sub_pixel_avg_variance64x64_avx2()
775 *sse = sse1 + sse2; in vpx_sub_pixel_avg_variance64x64_avx2()