| /external/libvpx/libvpx/vpx_dsp/x86/ |
| D | variance_sse2.c | 26 __m128i vsum = _mm_setzero_si128(); in vpx_get_mb_ss_sse2() local 56 static INLINE void variance_final_128_pel_sse2(__m128i vsse, __m128i vsum, in variance_final_128_pel_sse2() 68 static INLINE void variance_final_256_pel_sse2(__m128i vsse, __m128i vsum, in variance_final_256_pel_sse2() 80 static INLINE void variance_final_512_pel_sse2(__m128i vsse, __m128i vsum, in variance_final_512_pel_sse2() 220 __m128i vsse, vsum; in vpx_get8x8var_sse2() local 228 __m128i vsse, vsum; in vpx_get16x16var_sse2() local 236 __m128i vsse, vsum; in vpx_variance4x4_sse2() local 246 __m128i vsse, vsum; in vpx_variance4x8_sse2() local 256 __m128i vsse, vsum; in vpx_variance8x4_sse2() local 266 __m128i vsse, vsum; in vpx_variance8x8_sse2() local [all …]
|
| D | variance_avx2.c | 62 __m128i vsum, in variance_final_from_32bit_sum_avx2() 81 __m256i vsum, in variance_final_from_16bit_sum_avx2() 125 __m256i *const vsum) { in variance16_avx2() 140 __m256i *const vsum) { in variance32_avx2() 155 __m256i *const vsum) { in variance64_avx2() 170 __m256i vsse, vsum; in vpx_get16x16var_avx2() local 616 __m256i vsse, vsum; in vpx_variance16x8_avx2() local 626 __m256i vsse, vsum; in vpx_variance16x16_avx2() local 636 __m256i vsse, vsum; in vpx_variance16x32_avx2() local 646 __m256i vsse, vsum; in vpx_variance32x16_avx2() local [all …]
|
| /external/XNNPACK/src/u8-lut32norm/ |
| D | scalar.c | 23 uint32_t vsum = 0; in compute_sum() local 39 const uint32_t vsum = compute_sum(n, x, t); in xnn_u8_lut32norm_ukernel__scalar() local
|
| /external/XNNPACK/src/f32-gavgpool-spchw/ |
| D | neon-x4.c | 76 const float32x4_t vsum = vpaddq_f32(vsum01, vsum23); in xnn_f32_gavgpool_spchw_ukernel__neon_x4() local 82 const float32x4_t vsum = vcombine_f32(vpadd_f32(vget_low_f32(vsum01), vget_high_f32(vsum01)), in xnn_f32_gavgpool_spchw_ukernel__neon_x4() local 114 float32x2_t vsum = vadd_f32(vget_low_f32(vsum0), vget_high_f32(vsum0)); in xnn_f32_gavgpool_spchw_ukernel__neon_x4() local
|
| D | sse-x4.c | 78 const __m128 vsum = _mm_add_ps(_mm_movelh_ps(vsum01, vsum23), _mm_movehl_ps(vsum23, vsum01)); in xnn_f32_gavgpool_spchw_ukernel__sse_x4() local 94 __m128 vsum = _mm_setzero_ps(); in xnn_f32_gavgpool_spchw_ukernel__sse_x4() local
|
| /external/XNNPACK/src/f32-gavgpool/ |
| D | mp7p7q-neon.c | 54 const float32x4_t vsum = vaddq_f32(vsum016, vsum2345); in xnn_f32_gavgpool_ukernel_mp7p7q__neon() local 87 const float32x4_t vsum = vaddq_f32(vsum0123, vsum456a); in xnn_f32_gavgpool_ukernel_mp7p7q__neon() local 141 const float32x4_t vsum = vaddq_f32(vsum0123, vsum456a); in xnn_f32_gavgpool_ukernel_mp7p7q__neon() local 169 const float32x4_t vsum = vaddq_f32(vsum0123, vsum456a); in xnn_f32_gavgpool_ukernel_mp7p7q__neon() local
|
| D | mp7p7q-sse.c | 61 const __m128 vsum = _mm_add_ps(vsum016, vsum2345); in xnn_f32_gavgpool_ukernel_mp7p7q__sse() local 101 const __m128 vsum = _mm_add_ps(vsum0123, vsum456a); in xnn_f32_gavgpool_ukernel_mp7p7q__sse() local 163 const __m128 vsum = _mm_add_ps(vsum0123, vsum456a); in xnn_f32_gavgpool_ukernel_mp7p7q__sse() local 192 const __m128 vsum = _mm_add_ps(vsum0123, vsum456a); in xnn_f32_gavgpool_ukernel_mp7p7q__sse() local
|
| D | mp7p7q-psimd.c | 61 const psimd_f32 vsum = psimd_add_f32(vsum016, vsum2345); in xnn_f32_gavgpool_ukernel_mp7p7q__psimd() local 101 const psimd_f32 vsum = psimd_add_f32(vsum0123, vsum456a); in xnn_f32_gavgpool_ukernel_mp7p7q__psimd() local 163 const psimd_f32 vsum = psimd_add_f32(vsum0123, vsum456a); in xnn_f32_gavgpool_ukernel_mp7p7q__psimd() local 192 const psimd_f32 vsum = psimd_add_f32(vsum0123, vsum456a); in xnn_f32_gavgpool_ukernel_mp7p7q__psimd() local
|
| D | mp7p7q-scalar.c | 52 const float vsum = vsum016 + vsum2345; in xnn_f32_gavgpool_ukernel_mp7p7q__scalar() local 86 const float vsum = vsum0123 + vsum456a; in xnn_f32_gavgpool_ukernel_mp7p7q__scalar() local 140 const float vsum = vsum0123 + vsum456a; in xnn_f32_gavgpool_ukernel_mp7p7q__scalar() local
|
| D | mp7p7q-wasm.c | 52 const float vsum = vsum016 + vsum2345; in xnn_f32_gavgpool_ukernel_mp7p7q__wasm() local 86 const float vsum = vsum0123 + vsum456a; in xnn_f32_gavgpool_ukernel_mp7p7q__wasm() local 140 const float vsum = vsum0123 + vsum456a; in xnn_f32_gavgpool_ukernel_mp7p7q__wasm() local
|
| D | up7-psimd.c | 78 const psimd_f32 vsum = psimd_add_f32(vsum016, vsum2345); in xnn_f32_gavgpool_ukernel_up7__psimd() local 105 const psimd_f32 vsum = psimd_add_f32(vsum016, vsum2345); in xnn_f32_gavgpool_ukernel_up7__psimd() local
|
| D | up7-sse.c | 78 const __m128 vsum = _mm_add_ps(vsum016, vsum2345); in xnn_f32_gavgpool_ukernel_up7__sse() local 105 const __m128 vsum = _mm_add_ps(vsum016, vsum2345); in xnn_f32_gavgpool_ukernel_up7__sse() local
|
| D | up7-neon.c | 71 const float32x4_t vsum = vaddq_f32(vsum016, vsum2345); in xnn_f32_gavgpool_ukernel_up7__neon() local 97 const float32x4_t vsum = vaddq_f32(vsum016, vsum2345); in xnn_f32_gavgpool_ukernel_up7__neon() local
|
| /external/libaom/libaom/aom_dsp/x86/ |
| D | variance_avx2.c | 48 static INLINE int variance_final_from_32bit_sum_avx2(__m256i vsse, __m128i vsum, in variance_final_from_32bit_sum_avx2() 65 static INLINE int variance_final_512_avx2(__m256i vsse, __m256i vsum, in variance_final_512_avx2() 75 static INLINE int variance_final_1024_avx2(__m256i vsse, __m256i vsum, in variance_final_1024_avx2() 93 static INLINE int variance_final_2048_avx2(__m256i vsse, __m256i vsum, in variance_final_2048_avx2() 124 __m256i *const vsum) { in variance16_avx2() 137 __m256i *const vsum) { in variance32_avx2() 150 __m256i *const vsum) { in variance64_avx2() 164 __m256i *const vsum) { in variance128_avx2()
|
| /external/XNNPACK/src/q8-gavgpool/ |
| D | mp7p7q-scalar.c | 56 const uint32_t vsum = vsum016 + vsum2345; in xnn_q8_gavgpool_ukernel_mp7p7q__scalar() local 90 const uint32_t vsum = vsum016 + vsum2345; in xnn_q8_gavgpool_ukernel_mp7p7q__scalar() local 150 const uint32_t vsum = vsum016 + vsum2345; in xnn_q8_gavgpool_ukernel_mp7p7q__scalar() local
|
| D | mp7p7q-neon.c | 59 const int16x8_t vsum = vreinterpretq_s16_u16(vaddq_u16(vsum016, vsum2345)); in xnn_q8_gavgpool_ukernel_mp7p7q__neon() local 96 const int16x8_t vsum = vreinterpretq_s16_u16(vaddq_u16(vsum016, vsum2345)); in xnn_q8_gavgpool_ukernel_mp7p7q__neon() local 158 const int16x8_t vsum = vreinterpretq_s16_u16(vaddq_u16(vsum016, vsum2345)); in xnn_q8_gavgpool_ukernel_mp7p7q__neon() local 230 const int16x8_t vsum = vreinterpretq_s16_u16(vaddq_u16(vsum016, vsum2345)); in xnn_q8_gavgpool_ukernel_mp7p7q__neon() local
|
| /external/u-boot/cmd/ |
| D | md5sum.c | 43 static int parse_verify_sum(char *verify_str, u8 *vsum) in parse_verify_sum() 80 u8 vsum[16]; in do_md5sum() local
|
| /external/XNNPACK/src/f32-avgpool/ |
| D | mp9p8q-sse.c | 73 const __m128 vsum = _mm_add_ps(vsum2345, vsum01678); in xnn_f32_avgpool_ukernel_mp9p8q__sse() local 117 const __m128 vsum = _mm_add_ps(vsum2345, vsum0167a); in xnn_f32_avgpool_ukernel_mp9p8q__sse() local 184 const __m128 vsum = _mm_add_ps(vsum2345, vsum0167a); in xnn_f32_avgpool_ukernel_mp9p8q__sse() local 213 const __m128 vsum = _mm_add_ps(vsum2345, vsum0167a); in xnn_f32_avgpool_ukernel_mp9p8q__sse() local
|
| D | mp9p8q-neon.c | 64 const float32x4_t vsum = vaddq_f32(vsum2345, vsum01678); in xnn_f32_avgpool_ukernel_mp9p8q__neon() local 100 const float32x4_t vsum = vaddq_f32(vsum2345, vsum0167a); in xnn_f32_avgpool_ukernel_mp9p8q__neon() local 158 const float32x4_t vsum = vaddq_f32(vsum2345, vsum0167a); in xnn_f32_avgpool_ukernel_mp9p8q__neon() local 186 const float32x4_t vsum = vaddq_f32(vsum2345, vsum0167a); in xnn_f32_avgpool_ukernel_mp9p8q__neon() local
|
| D | mp9p8q-psimd.c | 73 const psimd_f32 vsum = psimd_add_f32(vsum2345, vsum01678); in xnn_f32_avgpool_ukernel_mp9p8q__psimd() local 118 const psimd_f32 vsum = psimd_add_f32(vsum2345, vsum0167a); in xnn_f32_avgpool_ukernel_mp9p8q__psimd() local 186 const psimd_f32 vsum = psimd_add_f32(vsum2345, vsum0167a); in xnn_f32_avgpool_ukernel_mp9p8q__psimd() local 215 const psimd_f32 vsum = psimd_add_f32(vsum2345, vsum0167a); in xnn_f32_avgpool_ukernel_mp9p8q__psimd() local
|
| D | mp9p8q-scalar.c | 64 const float vsum = vsum2345 + vsum01678; in xnn_f32_avgpool_ukernel_mp9p8q__scalar() local 101 const float vsum = vsum2345 + vsum0167a; in xnn_f32_avgpool_ukernel_mp9p8q__scalar() local 159 const float vsum = vsum2345 + vsum0167a; in xnn_f32_avgpool_ukernel_mp9p8q__scalar() local
|
| /external/XNNPACK/src/f32-pavgpool/ |
| D | mp9p8q-psimd.c | 73 const psimd_f32 vsum = psimd_add_f32(vsum2345, vsum01678); in xnn_f32_pavgpool_ukernel_mp9p8q__psimd() local 118 const psimd_f32 vsum = psimd_add_f32(vsum2345, vsum0167a); in xnn_f32_pavgpool_ukernel_mp9p8q__psimd() local 189 const psimd_f32 vsum = psimd_add_f32(vsum2345, vsum0167a); in xnn_f32_pavgpool_ukernel_mp9p8q__psimd() local 218 const psimd_f32 vsum = psimd_add_f32(vsum2345, vsum0167a); in xnn_f32_pavgpool_ukernel_mp9p8q__psimd() local
|
| D | mp9p8q-sse.c | 73 const __m128 vsum = _mm_add_ps(vsum2345, vsum01678); in xnn_f32_pavgpool_ukernel_mp9p8q__sse() local 117 const __m128 vsum = _mm_add_ps(vsum2345, vsum0167a); in xnn_f32_pavgpool_ukernel_mp9p8q__sse() local 187 const __m128 vsum = _mm_add_ps(vsum2345, vsum0167a); in xnn_f32_pavgpool_ukernel_mp9p8q__sse() local 216 const __m128 vsum = _mm_add_ps(vsum2345, vsum0167a); in xnn_f32_pavgpool_ukernel_mp9p8q__sse() local
|
| D | mp9p8q-neon.c | 64 const float32x4_t vsum = vaddq_f32(vsum2345, vsum01678); in xnn_f32_pavgpool_ukernel_mp9p8q__neon() local 100 const float32x4_t vsum = vaddq_f32(vsum2345, vsum0167a); in xnn_f32_pavgpool_ukernel_mp9p8q__neon() local 160 const float32x4_t vsum = vaddq_f32(vsum2345, vsum0167a); in xnn_f32_pavgpool_ukernel_mp9p8q__neon() local 188 const float32x4_t vsum = vaddq_f32(vsum2345, vsum0167a); in xnn_f32_pavgpool_ukernel_mp9p8q__neon() local
|
| D | mp9p8q-scalar.c | 64 const float vsum = vsum2345 + vsum01678; in xnn_f32_pavgpool_ukernel_mp9p8q__scalar() local 101 const float vsum = vsum2345 + vsum0167a; in xnn_f32_pavgpool_ukernel_mp9p8q__scalar() local 161 const float vsum = vsum2345 + vsum0167a; in xnn_f32_pavgpool_ukernel_mp9p8q__scalar() local
|