/external/XNNPACK/src/qs8-f32-vcvt/gen/ |
D | vcvt-avx-x16.c | 35 __m128i vx89AB = _mm_cvtepi8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qs8_f32_vcvt_ukernel__avx_x16() local
|
D | vcvt-sse41-x16.c | 35 __m128i vx89AB = _mm_cvtepi8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qs8_f32_vcvt_ukernel__sse41_x16() local
|
D | vcvt-avx-x24.c | 35 __m128i vx89AB = _mm_cvtepi8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qs8_f32_vcvt_ukernel__avx_x24() local
|
D | vcvt-sse41-x24.c | 35 __m128i vx89AB = _mm_cvtepi8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qs8_f32_vcvt_ukernel__sse41_x24() local
|
D | vcvt-sse41-x32.c | 35 __m128i vx89AB = _mm_cvtepi8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qs8_f32_vcvt_ukernel__sse41_x32() local
|
D | vcvt-avx-x32.c | 35 __m128i vx89AB = _mm_cvtepi8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qs8_f32_vcvt_ukernel__avx_x32() local
|
/external/XNNPACK/src/qu8-f32-vcvt/gen/ |
D | vcvt-avx-x16.c | 35 __m128i vx89AB = _mm_cvtepu8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qu8_f32_vcvt_ukernel__avx_x16() local
|
D | vcvt-sse41-x16.c | 35 __m128i vx89AB = _mm_cvtepu8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qu8_f32_vcvt_ukernel__sse41_x16() local
|
D | vcvt-avx-x24.c | 35 __m128i vx89AB = _mm_cvtepu8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qu8_f32_vcvt_ukernel__avx_x24() local
|
D | vcvt-sse41-x24.c | 35 __m128i vx89AB = _mm_cvtepu8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qu8_f32_vcvt_ukernel__sse41_x24() local
|
D | vcvt-sse41-x32.c | 35 __m128i vx89AB = _mm_cvtepu8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qu8_f32_vcvt_ukernel__sse41_x32() local
|
D | vcvt-avx-x32.c | 35 __m128i vx89AB = _mm_cvtepu8_epi32(_mm_cvtsi32_si128((int) unaligned_load_s32(x + 8))); in xnn_qu8_f32_vcvt_ukernel__avx_x32() local
|
/external/XNNPACK/src/f32-vhswish/gen/ |
D | vhswish-wasmsimd-x16.c | 35 v128_t vx89AB = wasm_v128_load(x + 8); in xnn_f32_vhswish_ukernel__wasmsimd_x16() local
|
D | vhswish-neon-x16.c | 35 float32x4_t vx89AB = vld1q_f32(x); x += 4; in xnn_f32_vhswish_ukernel__neon_x16() local
|
/external/XNNPACK/src/f32-qs8-vcvt/gen/ |
D | vcvt-sse41-x16.c | 38 __m128 vx89AB = _mm_loadu_ps(x + 8); in xnn_f32_qs8_vcvt_ukernel__sse41_x16() local
|
D | vcvt-sse2-x16.c | 38 __m128 vx89AB = _mm_loadu_ps(x + 8); in xnn_f32_qs8_vcvt_ukernel__sse2_x16() local
|
D | vcvt-neonv8-x16.c | 37 float32x4_t vx89AB = vld1q_f32(x); x += 4; in xnn_f32_qs8_vcvt_ukernel__neonv8_x16() local
|
D | vcvt-avx512skx-x64.c | 38 __m512 vx89AB = _mm512_loadu_ps(x + 32); in xnn_f32_qs8_vcvt_ukernel__avx512skx_x64() local
|
D | vcvt-neon-x16.c | 38 float32x4_t vx89AB = vld1q_f32(x); x += 4; in xnn_f32_qs8_vcvt_ukernel__neon_x16() local
|
D | vcvt-wasmsimd-cvt-x16.c | 37 v128_t vx89AB = wasm_v128_load(x + 8); in xnn_f32_qs8_vcvt_ukernel__wasmsimd_cvt_x16() local
|
/external/XNNPACK/src/f32-qu8-vcvt/gen/ |
D | vcvt-sse2-x16.c | 38 __m128 vx89AB = _mm_loadu_ps(x + 8); in xnn_f32_qu8_vcvt_ukernel__sse2_x16() local
|
D | vcvt-neonv8-x16.c | 37 float32x4_t vx89AB = vld1q_f32(x); x += 4; in xnn_f32_qu8_vcvt_ukernel__neonv8_x16() local
|
D | vcvt-avx512skx-x64.c | 38 __m512 vx89AB = _mm512_loadu_ps(x + 32); in xnn_f32_qu8_vcvt_ukernel__avx512skx_x64() local
|
D | vcvt-neon-x16.c | 38 float32x4_t vx89AB = vld1q_f32(x); x += 4; in xnn_f32_qu8_vcvt_ukernel__neon_x16() local
|
D | vcvt-wasmsimd-cvt-x16.c | 37 v128_t vx89AB = wasm_v128_load(x + 8); in xnn_f32_qu8_vcvt_ukernel__wasmsimd_cvt_x16() local
|