/external/XNNPACK/src/qs8-f32-vcvt/gen/ |
D | vcvt-sse41-x32.c | 40 __m128i vxSTUV = _mm_cvtepi8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 28))); in xnn_qs8_f32_vcvt_ukernel__sse41_x32() local
|
D | vcvt-avx-x32.c | 40 __m128i vxSTUV = _mm_cvtepi8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 28))); in xnn_qs8_f32_vcvt_ukernel__avx_x32() local
|
/external/XNNPACK/src/qu8-f32-vcvt/gen/ |
D | vcvt-sse41-x32.c | 40 __m128i vxSTUV = _mm_cvtepu8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 28))); in xnn_qu8_f32_vcvt_ukernel__sse41_x32() local
|
D | vcvt-avx-x32.c | 40 __m128i vxSTUV = _mm_cvtepu8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 28))); in xnn_qu8_f32_vcvt_ukernel__avx_x32() local
|
/external/XNNPACK/src/f32-qs8-vcvt/gen/ |
D | vcvt-sse2-x32.c | 43 __m128 vxSTUV = _mm_loadu_ps(x + 28); in xnn_f32_qs8_vcvt_ukernel__sse2_x32() local
|
D | vcvt-sse41-x32.c | 43 __m128 vxSTUV = _mm_loadu_ps(x + 28); in xnn_f32_qs8_vcvt_ukernel__sse41_x32() local
|
D | vcvt-avx512skx-x128.c | 43 __m512 vxSTUV = _mm512_loadu_ps(x + 112); in xnn_f32_qs8_vcvt_ukernel__avx512skx_x128() local
|
D | vcvt-neon-x32.c | 43 float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_qs8_vcvt_ukernel__neon_x32() local
|
D | vcvt-wasmsimd-cvt-x32.c | 42 v128_t vxSTUV = wasm_v128_load(x + 28); in xnn_f32_qs8_vcvt_ukernel__wasmsimd_cvt_x32() local
|
D | vcvt-neonv8-x32.c | 42 float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_qs8_vcvt_ukernel__neonv8_x32() local
|
D | vcvt-wasmsimd-magic-x32.c | 43 v128_t vxSTUV = wasm_v128_load(x + 28); in xnn_f32_qs8_vcvt_ukernel__wasmsimd_magic_x32() local
|
/external/XNNPACK/src/f32-qu8-vcvt/gen/ |
D | vcvt-sse2-x32.c | 43 __m128 vxSTUV = _mm_loadu_ps(x + 28); in xnn_f32_qu8_vcvt_ukernel__sse2_x32() local
|
D | vcvt-avx512skx-x128.c | 43 __m512 vxSTUV = _mm512_loadu_ps(x + 112); in xnn_f32_qu8_vcvt_ukernel__avx512skx_x128() local
|
D | vcvt-neonv8-x32.c | 42 float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_qu8_vcvt_ukernel__neonv8_x32() local
|
D | vcvt-wasmsimd-cvt-x32.c | 42 v128_t vxSTUV = wasm_v128_load(x + 28); in xnn_f32_qu8_vcvt_ukernel__wasmsimd_cvt_x32() local
|
D | vcvt-neon-x32.c | 43 float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_qu8_vcvt_ukernel__neon_x32() local
|
D | vcvt-wasmsimd-magic-x32.c | 43 v128_t vxSTUV = wasm_v128_load(x + 28); in xnn_f32_qu8_vcvt_ukernel__wasmsimd_magic_x32() local
|
/external/XNNPACK/src/f32-vsqrt/gen/ |
D | neonfma-nr2fma1adj-x32.c | 37 const float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_vsqrt_ukernel__neonfma_nr2fma1adj_x32() local
|
D | neonfma-nr2fma1adj-x36.c | 37 const float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_vsqrt_ukernel__neonfma_nr2fma1adj_x36() local
|
D | neonfma-nr2fma1adj-x40.c | 37 const float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_vsqrt_ukernel__neonfma_nr2fma1adj_x40() local
|
D | neonfma-nr1rsqrts1fma1adj-x32.c | 37 const float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_vsqrt_ukernel__neonfma_nr1rsqrts1fma1adj_x32() local
|
D | neonfma-nr1rsqrts1fma1adj-x36.c | 37 const float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_vsqrt_ukernel__neonfma_nr1rsqrts1fma1adj_x36() local
|
D | neonfma-nr1rsqrts1fma1adj-x40.c | 37 const float32x4_t vxSTUV = vld1q_f32(x); x += 4; in xnn_f32_vsqrt_ukernel__neonfma_nr1rsqrts1fma1adj_x40() local
|
/external/XNNPACK/src/amalgam/ |
D | avx512skx.c | 111 __m512 vxSTUV = _mm512_loadu_ps(x + 112); in xnn_f32_qs8_vcvt_ukernel__avx512skx_x128() local 229 __m512 vxSTUV = _mm512_loadu_ps(x + 112); in xnn_f32_qu8_vcvt_ukernel__avx512skx_x128() local
|
D | avx.c | 8273 __m128i vxSTUV = _mm_cvtepi8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 28))); in xnn_qs8_f32_vcvt_ukernel__avx_x32() local 11004 __m128i vxSTUV = _mm_cvtepu8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 28))); in xnn_qu8_f32_vcvt_ukernel__avx_x32() local
|