/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/x86/ |
D | vp9_variance_sse2.c | 143 int avg; in vp9_variance4x4_sse2() local 146 &var, &avg, vp9_get4x4var_mmx, 4); in vp9_variance4x4_sse2() 148 return (var - (((unsigned int)avg * avg) >> 4)); in vp9_variance4x4_sse2() 157 int avg; in vp9_variance8x4_sse2() local 160 &var, &avg, vp9_get4x4var_mmx, 4); in vp9_variance8x4_sse2() 162 return (var - (((unsigned int)avg * avg) >> 5)); in vp9_variance8x4_sse2() 171 int avg; in vp9_variance4x8_sse2() local 174 &var, &avg, vp9_get4x4var_mmx, 4); in vp9_variance4x8_sse2() 176 return (var - (((unsigned int)avg * avg) >> 5)); in vp9_variance4x8_sse2() 187 int avg; in vp9_variance8x8_sse2() local [all …]
|
D | vp9_variance_mmx.c | 42 int avg; in vp9_variance4x4_mmx() local 44 vp9_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg); in vp9_variance4x4_mmx() 46 return (var - (((unsigned int)avg * avg) >> 4)); in vp9_variance4x4_mmx() 56 int avg; in vp9_variance8x8_mmx() local 58 vp9_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg); in vp9_variance8x8_mmx() 61 return (var - (((unsigned int)avg * avg) >> 6)); in vp9_variance8x8_mmx() 96 int sum0, sum1, sum2, sum3, avg; in vp9_variance16x16_mmx() local 108 avg = sum0 + sum1 + sum2 + sum3; in vp9_variance16x16_mmx() 110 return (var - (((unsigned int)avg * avg) >> 8)); in vp9_variance16x16_mmx() 120 int sum0, sum1, avg; in vp9_variance16x8_mmx() local [all …]
|
D | vp9_variance_avx2.c | 101 int avg; in vp9_variance16x16_avx2() local 104 &var, &avg, vp9_get16x16var_avx2, 16); in vp9_variance16x16_avx2() 106 return (var - (((unsigned int)avg * avg) >> 8)); in vp9_variance16x16_avx2() 129 int avg; in vp9_variance32x32_avx2() local 133 &var, &avg, vp9_get32x32var_avx2, 32); in vp9_variance32x32_avx2() 135 return (var - (((int64_t)avg * avg) >> 10)); in vp9_variance32x32_avx2() 144 int avg; in vp9_variance32x16_avx2() local 148 &var, &avg, vp9_get32x32var_avx2, 32); in vp9_variance32x16_avx2() 150 return (var - (((int64_t)avg * avg) >> 9)); in vp9_variance32x16_avx2() 160 int avg; in vp9_variance64x64_avx2() local [all …]
|
D | vp9_subpel_variance.asm | 142 %if %2 == 1 ; avg 156 %if %2 == 1 ; avg 195 %if %2 == 1 ; avg 433 %if %2 == 1 ; avg 477 %if %2 == 1 ; avg 490 %if %2 == 1 ; avg 515 %if %2 == 1 ; avg 528 %if %2 == 1 ; avg 544 %else ; !avg 561 %if %2 == 1 ; avg [all …]
|
D | vp9_subpel_variance_impl_sse2.asm | 47 … pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) horizontal line 1 54 … pavgb xmm1, xmm2 ; xmm1 = avg(xmm1,xmm3) horizontal line i+1 165 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) 271 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3)
|
D | vp9_sad_sse2.asm | 23 %else ; avg 38 %endif ; avg/sad
|
D | vp9_variance_impl_sse2.asm | 439 … pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) horizontal line 1 451 … pavgb xmm1, xmm2 ; xmm1 = avg(xmm1,xmm3) horizontal line i+1 562 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) 669 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3)
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/ |
D | variance_c.c | 73 int avg; in vp8_variance16x16_c() local 76 variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 16, &var, &avg); in vp8_variance16x16_c() 78 return (var - (((unsigned int)avg * avg) >> 8)); in vp8_variance16x16_c() 89 int avg; in vp8_variance8x16_c() local 92 variance(src_ptr, source_stride, ref_ptr, recon_stride, 8, 16, &var, &avg); in vp8_variance8x16_c() 94 return (var - (((unsigned int)avg * avg) >> 7)); in vp8_variance8x16_c() 105 int avg; in vp8_variance16x8_c() local 108 variance(src_ptr, source_stride, ref_ptr, recon_stride, 16, 8, &var, &avg); in vp8_variance16x8_c() 110 return (var - (((unsigned int)avg * avg) >> 7)); in vp8_variance16x8_c() 122 int avg; in vp8_variance8x8_c() local [all …]
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/encoder/ |
D | vp9_variance.c | 172 int avg; in vp9_variance64x32_c() local 174 variance(src_ptr, source_stride, ref_ptr, recon_stride, 64, 32, &var, &avg); in vp9_variance64x32_c() 176 return (var - (((int64_t)avg * avg) >> 11)); in vp9_variance64x32_c() 229 int avg; in vp9_variance32x64_c() local 231 variance(src_ptr, source_stride, ref_ptr, recon_stride, 32, 64, &var, &avg); in vp9_variance32x64_c() 233 return (var - (((int64_t)avg * avg) >> 11)); in vp9_variance32x64_c() 286 int avg; in vp9_variance32x16_c() local 288 variance(src_ptr, source_stride, ref_ptr, recon_stride, 32, 16, &var, &avg); in vp9_variance32x16_c() 290 return (var - (((int64_t)avg * avg) >> 9)); in vp9_variance32x16_c() 343 int avg; in vp9_variance16x32_c() local [all …]
|
D | vp9_aq_variance.c | 127 int avg; in block_variance() local 129 vp9_64_zeros, 0, bw, bh, &sse, &avg); in block_variance() 130 var = sse - (((int64_t)avg * avg) / (bw * bh)); in block_variance()
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp8/common/x86/ |
D | variance_mmx.c | 90 int avg; in vp8_variance4x4_mmx() local 92 vp8_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg) ; in vp8_variance4x4_mmx() 94 return (var - (((unsigned int)avg * avg) >> 4)); in vp8_variance4x4_mmx() 106 int avg; in vp8_variance8x8_mmx() local 108 vp8_get8x8var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg) ; in vp8_variance8x8_mmx() 111 return (var - (((unsigned int)avg * avg) >> 6)); in vp8_variance8x8_mmx() 145 int sum0, sum1, sum2, sum3, avg; in vp8_variance16x16_mmx() local 154 avg = sum0 + sum1 + sum2 + sum3; in vp8_variance16x16_mmx() 156 return (var - (((unsigned int)avg * avg) >> 8)); in vp8_variance16x16_mmx() 167 int sum0, sum1, avg; in vp8_variance16x8_mmx() local [all …]
|
D | variance_sse2.c | 147 int avg; in vp8_variance4x4_wmt() local 149 vp8_get4x4var_mmx(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg) ; in vp8_variance4x4_wmt() 151 return (var - (((unsigned int)avg * avg) >> 4)); in vp8_variance4x4_wmt() 164 int avg; in vp8_variance8x8_wmt() local 166 vp8_get8x8var_sse2(src_ptr, source_stride, ref_ptr, recon_stride, &var, &avg) ; in vp8_variance8x8_wmt() 168 return (var - (((unsigned int)avg * avg) >> 6)); in vp8_variance8x8_wmt() 215 int sum0, sum1, avg; in vp8_variance16x8_wmt() local 221 avg = sum0 + sum1; in vp8_variance16x8_wmt() 223 return (var - (((unsigned int)avg * avg) >> 7)); in vp8_variance16x8_wmt() 236 int sum0, sum1, avg; in vp8_variance8x16_wmt() local [all …]
|
D | variance_impl_sse2.asm | 721 … pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) horizontal line 1 733 … pavgb xmm1, xmm2 ; xmm1 = avg(xmm1,xmm3) horizontal line i+1 839 … pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) horizontal line 1 846 … pavgb xmm1, xmm2 ; xmm1 = avg(xmm1,xmm3) horizontal line i+1 960 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) 1065 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) 1174 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3) 1277 pavgb xmm5, xmm3 ; xmm5 = avg(xmm1,xmm3)
|
/hardware/intel/common/omx-components/videocodec/libvpx_internal/libvpx/vp9/common/x86/ |
D | vp9_asm_stubs.c | 26 #define FUN_CONV_1D(name, step_q4, filter, dir, src_start, avg, opt) \ argument 35 vp9_filter_block1d16_##dir##8_##avg##opt(src_start, \ 46 vp9_filter_block1d8_##dir##8_##avg##opt(src_start, \ 57 vp9_filter_block1d4_##dir##8_##avg##opt(src_start, \ 69 vp9_filter_block1d16_##dir##2_##avg##opt(src, \ 80 vp9_filter_block1d8_##dir##2_##avg##opt(src, \ 91 vp9_filter_block1d4_##dir##2_##avg##opt(src, \ 110 #define FUN_CONV_2D(avg, opt) \ argument 111 void vp9_convolve8_##avg##opt(const uint8_t *src, ptrdiff_t src_stride, \ 125 vp9_convolve8_##avg##vert_##opt(fdata2 + 3 * 64, 64, dst, dst_stride, \ [all …]
|
D | vp9_copy_sse2.asm | 36 %ifidn %1, avg 59 %ifidn %1, avg 84 %ifidn %1, avg 110 %ifidn %1, avg 135 %ifidn %1, avg 152 convolve_fn avg
|
D | vp9_intrapred_ssse3.asm | 310 ; result = avg(x,z) 312 ; result = avg(result,y) 469 X_PLUS_2Y_PLUS_Z_PLUS_2_RSH_2 m0, m1, m2, m3 ; 3-tap avg B4 B3 B2 B1 C1 D1 470 pavgb m1, m0 ; 2-tap avg A4 A3 A2 A1 507 pavgb m6, m1, m2 ; 2-tap avg A8-A1 509 X_PLUS_2Y_PLUS_Z_PLUS_2_RSH_2 m0, m4, m5, m7 ; 3-tap avg C-H1 511 X_PLUS_2Y_PLUS_Z_PLUS_2_RSH_2 m1, m2, m3, m0 ; 3-tap avg B8-1 562 X_PLUS_2Y_PLUS_Z_PLUS_2_RSH_2 m0, m5, m3, m4 ; 3-tap avg B3-Bg 572 X_PLUS_2Y_PLUS_Z_PLUS_2_RSH_2 m7, m3, m5, m1 ; 3-tap avg C1-P1 625 X_PLUS_2Y_PLUS_Z_PLUS_2_RSH_2 m1, m4, m6, m2 ; 3-tap avg above [high] [all …]
|
/hardware/invensense/60xx/mlsdk/mllite/ |
D | compass.c | 107 float avg, sum, median, sorted[YAS_DEFAULT_FILTER_LEN]; in adaptive_filter_filter() local 123 avg = 0; in adaptive_filter_filter() 125 avg += adap_filter->sequence[i]; in adaptive_filter_filter() 127 avg /= adap_filter->len; in adaptive_filter_filter() 135 sum += square(avg - adap_filter->sequence[i]); in adaptive_filter_filter() 143 return ((in - avg) * (sum - adap_filter->noise) / sum + avg); in adaptive_filter_filter()
|