/external/XNNPACK/src/f32-f16-vcvt/gen/ |
D | vcvt-avx-x8.c | 50 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x8() local 100 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x8() local
|
D | vcvt-sse2-x8.c | 50 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x8() local 100 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x8() local
|
D | vcvt-wasmsimd-x8.c | 49 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x8() local 99 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x8() local
|
D | vcvt-sse41-x8.c | 50 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local 100 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local
|
D | vcvt-sse41-x16.c | 143 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x16() local 193 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x16() local
|
D | vcvt-sse2-x16.c | 144 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local 194 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local
|
D | vcvt-avx-x16.c | 143 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x16() local 193 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x16() local
|
D | vcvt-wasmsimd-x16.c | 140 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x16() local 190 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x16() local
|
D | vcvt-sse2-x24.c | 179 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x24() local 229 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x24() local
|
D | vcvt-wasmsimd-x24.c | 174 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x24() local 224 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x24() local
|
D | vcvt-sse41-x24.c | 177 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x24() local 227 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x24() local
|
D | vcvt-avx-x24.c | 177 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x24() local 227 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x24() local
|
D | vcvt-sse2-x32.c | 214 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x32() local 264 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x32() local
|
D | vcvt-sse41-x32.c | 211 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x32() local 261 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x32() local
|
D | vcvt-wasmsimd-x32.c | 208 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x32() local 258 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x32() local
|
D | vcvt-avx-x32.c | 211 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x32() local 261 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x32() local
|
/external/XNNPACK/src/f32-f16-vcvt/ |
D | wasmsimd.c.in | 114 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); variable 164 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); variable
|
/external/XNNPACK/src/math/ |
D | cvt-f32-f16-neon.c | 40 uint32x4_t vbias_lo = vaddq_u32(vreinterpretq_u32_f32(vabsx_lo), vexp_bias); in xnn_math_f32_f16_cvt__neon() local
|
D | cvt-f32-f16-sse2.c | 43 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_math_f32_f16_cvt__sse2() local
|
D | cvt-f32-f16-wasmsimd.c | 42 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_math_f32_f16_cvt__wasmsimd() local
|
D | cvt-f32-f16-sse41.c | 43 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_math_f32_f16_cvt__sse41() local
|
/external/XNNPACK/src/amalgam/ |
D | sse41.c | 190 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local 240 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local
|
D | avx.c | 1391 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x24() local 1441 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x24() local
|
D | sse2.c | 1014 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local 1064 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local
|