| /external/XNNPACK/src/qu8-vmulc/gen/ |
| D | minmax-fp32-neonv8-ld64-x16.c | 49 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qu8_vmulc_minmax_fp32_ukernel__neonv8_ld64_x16() local
|
| D | minmax-fp32-neon-ld64-x16.c | 49 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qu8_vmulc_minmax_fp32_ukernel__neon_ld64_x16() local
|
| D | minmax-fp32-neonv8-ld128-x16.c | 57 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qu8_vmulc_minmax_fp32_ukernel__neonv8_ld128_x16() local
|
| D | minmax-fp32-neon-ld128-x16.c | 57 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qu8_vmulc_minmax_fp32_ukernel__neon_ld128_x16() local
|
| D | minmax-fp32-avx-mul16-ld64-x16.c | 57 __m128 vfpaccCDEF = _mm_cvtepi32_ps(vprodCDEF); in xnn_qu8_vmulc_minmax_fp32_ukernel__avx_mul16_ld64_x16() local
|
| D | minmax-fp32-sse41-mul16-ld64-x16.c | 57 __m128 vfpaccCDEF = _mm_cvtepi32_ps(vprodCDEF); in xnn_qu8_vmulc_minmax_fp32_ukernel__sse41_mul16_ld64_x16() local
|
| D | minmax-fp32-sse2-mul16-ld64-x16.c | 60 __m128 vfpaccCDEF = _mm_cvtepi32_ps(vprodCDEF); in xnn_qu8_vmulc_minmax_fp32_ukernel__sse2_mul16_ld64_x16() local
|
| /external/XNNPACK/src/qs8-vmulc/gen/ |
| D | minmax-fp32-neonv8-ld64-x16.c | 49 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qs8_vmulc_minmax_fp32_ukernel__neonv8_ld64_x16() local
|
| D | minmax-fp32-neon-ld64-x16.c | 49 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qs8_vmulc_minmax_fp32_ukernel__neon_ld64_x16() local
|
| D | minmax-fp32-neonv8-ld128-x16.c | 57 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qs8_vmulc_minmax_fp32_ukernel__neonv8_ld128_x16() local
|
| D | minmax-fp32-neon-ld128-x16.c | 57 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qs8_vmulc_minmax_fp32_ukernel__neon_ld128_x16() local
|
| D | minmax-fp32-avx-mul16-ld64-x16.c | 57 __m128 vfpaccCDEF = _mm_cvtepi32_ps(vprodCDEF); in xnn_qs8_vmulc_minmax_fp32_ukernel__avx_mul16_ld64_x16() local
|
| D | minmax-fp32-sse41-mul16-ld64-x16.c | 57 __m128 vfpaccCDEF = _mm_cvtepi32_ps(vprodCDEF); in xnn_qs8_vmulc_minmax_fp32_ukernel__sse41_mul16_ld64_x16() local
|
| D | minmax-fp32-sse2-mul16-ld64-x16.c | 59 __m128 vfpaccCDEF = _mm_cvtepi32_ps(vprodCDEF); in xnn_qs8_vmulc_minmax_fp32_ukernel__sse2_mul16_ld64_x16() local
|
| /external/XNNPACK/src/qu8-vmul/gen/ |
| D | minmax-fp32-neonv8-ld64-x16.c | 51 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qu8_vmul_minmax_fp32_ukernel__neonv8_ld64_x16() local
|
| D | minmax-fp32-neon-ld64-x16.c | 51 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qu8_vmul_minmax_fp32_ukernel__neon_ld64_x16() local
|
| D | minmax-fp32-neonv8-ld128-x16.c | 61 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qu8_vmul_minmax_fp32_ukernel__neonv8_ld128_x16() local
|
| D | minmax-fp32-neon-ld128-x16.c | 61 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qu8_vmul_minmax_fp32_ukernel__neon_ld128_x16() local
|
| /external/XNNPACK/src/qs8-vmul/gen/ |
| D | minmax-fp32-neonv8-ld64-x16.c | 51 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qs8_vmul_minmax_fp32_ukernel__neonv8_ld64_x16() local
|
| D | minmax-fp32-neon-ld64-x16.c | 51 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qs8_vmul_minmax_fp32_ukernel__neon_ld64_x16() local
|
| D | minmax-fp32-neonv8-ld128-x16.c | 61 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qs8_vmul_minmax_fp32_ukernel__neonv8_ld128_x16() local
|
| D | minmax-fp32-neon-ld128-x16.c | 61 float32x4_t vfpaccCDEF = vcvtq_f32_s32(vaccCDEF); in xnn_qs8_vmul_minmax_fp32_ukernel__neon_ld128_x16() local
|
| D | minmax-fp32-sse41-mul16-ld64-x16.c | 60 __m128 vfpaccCDEF = _mm_cvtepi32_ps(vprodCDEF); in xnn_qs8_vmul_minmax_fp32_ukernel__sse41_mul16_ld64_x16() local
|
| D | minmax-fp32-avx-mul16-ld64-x16.c | 60 __m128 vfpaccCDEF = _mm_cvtepi32_ps(vprodCDEF); in xnn_qs8_vmul_minmax_fp32_ukernel__avx_mul16_ld64_x16() local
|
| D | minmax-fp32-sse2-mul16-ld64-x16.c | 64 __m128 vfpaccCDEF = _mm_cvtepi32_ps(vprodCDEF); in xnn_qs8_vmul_minmax_fp32_ukernel__sse2_mul16_ld64_x16() local
|