/external/XNNPACK/src/f32-pavgpool/ |
D | up9-psimd.c | 93 const psimd_f32 vsum018 = psimd_add_f32(psimd_add_f32(vi0, vi1), vi8); in xnn_f32_pavgpool_ukernel_up9__psimd() local 99 const psimd_f32 vsum01678 = psimd_add_f32(vsum018, vsum67); in xnn_f32_pavgpool_ukernel_up9__psimd() 126 const psimd_f32 vsum018 = psimd_add_f32(vsum01, vi8); in xnn_f32_pavgpool_ukernel_up9__psimd() local 128 const psimd_f32 vsum01678 = psimd_add_f32(vsum018, vsum67); in xnn_f32_pavgpool_ukernel_up9__psimd()
|
D | up9-sse.c | 93 const __m128 vsum018 = _mm_add_ps(_mm_add_ps(vi0, vi1), vi8); in xnn_f32_pavgpool_ukernel_up9__sse() local 99 const __m128 vsum01678 = _mm_add_ps(vsum018, vsum67); in xnn_f32_pavgpool_ukernel_up9__sse() 125 const __m128 vsum018 = _mm_add_ps(vsum01, vi8); in xnn_f32_pavgpool_ukernel_up9__sse() local 127 const __m128 vsum01678 = _mm_add_ps(vsum018, vsum67); in xnn_f32_pavgpool_ukernel_up9__sse()
|
D | up9-neon.c | 87 const float32x4_t vsum018 = vaddq_f32(vsum01, vi8); in xnn_f32_pavgpool_ukernel_up9__neon() local 89 const float32x4_t vsum01678 = vaddq_f32(vsum018, vsum67); in xnn_f32_pavgpool_ukernel_up9__neon() 115 const float32x4_t vsum018 = vaddq_f32(vsum01, vi8); in xnn_f32_pavgpool_ukernel_up9__neon() local 117 const float32x4_t vsum01678 = vaddq_f32(vsum018, vsum67); in xnn_f32_pavgpool_ukernel_up9__neon()
|
D | up9-scalar.c | 86 const float vsum018 = vsum01 + vi8; in xnn_f32_pavgpool_ukernel_up9__scalar() local 88 const float vsum01678 = vsum018 + vsum67; in xnn_f32_pavgpool_ukernel_up9__scalar()
|
D | up9-wasm.c | 86 const float vsum018 = vsum01 + vi8; in xnn_f32_pavgpool_ukernel_up9__wasm() local 88 const float vsum01678 = vsum018 + vsum67; in xnn_f32_pavgpool_ukernel_up9__wasm()
|
D | mp9p8q-scalar.c | 61 const float vsum018 = vsum01 + vi8; in xnn_f32_pavgpool_ukernel_mp9p8q__scalar() local 63 const float vsum01678 = vsum018 + vsum67; in xnn_f32_pavgpool_ukernel_mp9p8q__scalar()
|
D | mp9p8q-wasm.c | 61 const float vsum018 = vsum01 + vi8; in xnn_f32_pavgpool_ukernel_mp9p8q__wasm() local 63 const float vsum01678 = vsum018 + vsum67; in xnn_f32_pavgpool_ukernel_mp9p8q__wasm()
|
D | mp9p8q-psimd.c | 70 const psimd_f32 vsum018 = psimd_add_f32(vsum01, vi8); in xnn_f32_pavgpool_ukernel_mp9p8q__psimd() local 72 const psimd_f32 vsum01678 = psimd_add_f32(vsum018, vsum67); in xnn_f32_pavgpool_ukernel_mp9p8q__psimd()
|
D | mp9p8q-sse.c | 70 const __m128 vsum018 = _mm_add_ps(vsum01, vi8); in xnn_f32_pavgpool_ukernel_mp9p8q__sse() local 72 const __m128 vsum01678 = _mm_add_ps(vsum018, vsum67); in xnn_f32_pavgpool_ukernel_mp9p8q__sse()
|
D | mp9p8q-neon.c | 61 const float32x4_t vsum018 = vaddq_f32(vsum01, vi8); in xnn_f32_pavgpool_ukernel_mp9p8q__neon() local 63 const float32x4_t vsum01678 = vaddq_f32(vsum018, vsum67); in xnn_f32_pavgpool_ukernel_mp9p8q__neon()
|
/external/XNNPACK/src/f32-avgpool/ |
D | up9-neon.c | 85 const float32x4_t vsum018 = vaddq_f32(vsum01, vi8); in xnn_f32_avgpool_ukernel_up9__neon() local 87 const float32x4_t vsum01678 = vaddq_f32(vsum018, vsum67); in xnn_f32_avgpool_ukernel_up9__neon() 113 const float32x4_t vsum018 = vaddq_f32(vsum01, vi8); in xnn_f32_avgpool_ukernel_up9__neon() local 115 const float32x4_t vsum01678 = vaddq_f32(vsum018, vsum67); in xnn_f32_avgpool_ukernel_up9__neon()
|
D | up9-sse.c | 90 const __m128 vsum018 = _mm_add_ps(_mm_add_ps(vi0, vi1), vi8); in xnn_f32_avgpool_ukernel_up9__sse() local 96 const __m128 vsum01678 = _mm_add_ps(vsum018, vsum67); in xnn_f32_avgpool_ukernel_up9__sse() 122 const __m128 vsum018 = _mm_add_ps(vsum01, vi8); in xnn_f32_avgpool_ukernel_up9__sse() local 124 const __m128 vsum01678 = _mm_add_ps(vsum018, vsum67); in xnn_f32_avgpool_ukernel_up9__sse()
|
D | up9-psimd.c | 90 const psimd_f32 vsum018 = psimd_add_f32(psimd_add_f32(vi0, vi1), vi8); in xnn_f32_avgpool_ukernel_up9__psimd() local 96 const psimd_f32 vsum01678 = psimd_add_f32(vsum018, vsum67); in xnn_f32_avgpool_ukernel_up9__psimd() 123 const psimd_f32 vsum018 = psimd_add_f32(vsum01, vi8); in xnn_f32_avgpool_ukernel_up9__psimd() local 125 const psimd_f32 vsum01678 = psimd_add_f32(vsum018, vsum67); in xnn_f32_avgpool_ukernel_up9__psimd()
|
D | up9-scalar.c | 84 const float vsum018 = vsum01 + vi8; in xnn_f32_avgpool_ukernel_up9__scalar() local 86 const float vsum01678 = vsum018 + vsum67; in xnn_f32_avgpool_ukernel_up9__scalar()
|
D | up9-wasm.c | 84 const float vsum018 = vsum01 + vi8; in xnn_f32_avgpool_ukernel_up9__wasm() local 86 const float vsum01678 = vsum018 + vsum67; in xnn_f32_avgpool_ukernel_up9__wasm()
|
D | mp9p8q-scalar.c | 61 const float vsum018 = vsum01 + vi8; in xnn_f32_avgpool_ukernel_mp9p8q__scalar() local 63 const float vsum01678 = vsum018 + vsum67; in xnn_f32_avgpool_ukernel_mp9p8q__scalar()
|
D | mp9p8q-wasm.c | 61 const float vsum018 = vsum01 + vi8; in xnn_f32_avgpool_ukernel_mp9p8q__wasm() local 63 const float vsum01678 = vsum018 + vsum67; in xnn_f32_avgpool_ukernel_mp9p8q__wasm()
|
D | mp9p8q-sse.c | 70 const __m128 vsum018 = _mm_add_ps(vsum01, vi8); in xnn_f32_avgpool_ukernel_mp9p8q__sse() local 72 const __m128 vsum01678 = _mm_add_ps(vsum018, vsum67); in xnn_f32_avgpool_ukernel_mp9p8q__sse()
|
D | mp9p8q-neon.c | 61 const float32x4_t vsum018 = vaddq_f32(vsum01, vi8); in xnn_f32_avgpool_ukernel_mp9p8q__neon() local 63 const float32x4_t vsum01678 = vaddq_f32(vsum018, vsum67); in xnn_f32_avgpool_ukernel_mp9p8q__neon()
|
D | mp9p8q-psimd.c | 70 const psimd_f32 vsum018 = psimd_add_f32(vsum01, vi8); in xnn_f32_avgpool_ukernel_mp9p8q__psimd() local 72 const psimd_f32 vsum01678 = psimd_add_f32(vsum018, vsum67); in xnn_f32_avgpool_ukernel_mp9p8q__psimd()
|
/external/XNNPACK/src/q8-avgpool/ |
D | up9-neon.c | 92 const uint16x8_t vsum018 = vaddw_u8(vaddl_u8(vi0, vi1), vi8); in xnn_q8_avgpool_ukernel_up9__neon() local 98 const uint16x8_t vsum01678 = vaddq_u16(vsum018, vsum67); in xnn_q8_avgpool_ukernel_up9__neon() 165 const uint16x8_t vsum018 = vaddw_u8(vaddl_u8(vi0, vi1), vi8); in xnn_q8_avgpool_ukernel_up9__neon() local 171 const uint16x8_t vsum01678 = vaddq_u16(vsum018, vsum67); in xnn_q8_avgpool_ukernel_up9__neon()
|
D | up9-scalar.c | 90 const uint32_t vsum018 = vsum01 + vi8; in xnn_q8_avgpool_ukernel_up9__scalar() local 92 const uint32_t vsum01678 = vsum018 + vsum67; in xnn_q8_avgpool_ukernel_up9__scalar()
|
D | up9-sse2.c | 96 const __m128i vsum018 = _mm_add_epi16(_mm_add_epi16(vxi0, vxi1), vxi8); in xnn_q8_avgpool_ukernel_up9__sse2() local 102 const __m128i vsum01678 = _mm_add_epi16(vsum018, vsum67); in xnn_q8_avgpool_ukernel_up9__sse2() 171 const __m128i vsum018 = _mm_add_epi16(_mm_add_epi16(vxi0, vxi1), vxi8); in xnn_q8_avgpool_ukernel_up9__sse2() local 177 const __m128i vsum01678 = _mm_add_epi16(vsum018, vsum67); in xnn_q8_avgpool_ukernel_up9__sse2()
|
D | mp9p8q-scalar.c | 68 const uint32_t vsum018 = vsum01 + vi8; in xnn_q8_avgpool_ukernel_mp9p8q__scalar() local 70 const uint32_t vsum01678 = vsum018 + vsum67; in xnn_q8_avgpool_ukernel_mp9p8q__scalar()
|
D | mp9p8q-neon.c | 68 const uint16x8_t vsum018 = vaddw_u8(vaddl_u8(vi0, vi1), vi8); in xnn_q8_avgpool_ukernel_mp9p8q__neon() local 74 const uint16x8_t vsum01678 = vaddq_u16(vsum018, vsum67); in xnn_q8_avgpool_ukernel_mp9p8q__neon()
|