| /external/XNNPACK/src/f32-f16-vcvt/gen/ |
| D | vcvt-sse2-x8.c | 51 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x8() local 101 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x8() local
|
| D | vcvt-avx-x8.c | 51 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x8() local 101 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x8() local
|
| D | vcvt-sse41-x8.c | 51 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local 101 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local
|
| D | vcvt-wasmsimd-x8.c | 49 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x8() local 99 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x8() local
|
| D | vcvt-avx-x16.c | 144 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x16() local 194 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x16() local
|
| D | vcvt-wasmsimd-x16.c | 140 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x16() local 190 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x16() local
|
| D | vcvt-sse41-x16.c | 144 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x16() local 194 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x16() local
|
| D | vcvt-sse2-x16.c | 145 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local 195 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local
|
| D | vcvt-sse2-x24.c | 180 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x24() local 230 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x24() local
|
| D | vcvt-wasmsimd-x24.c | 174 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x24() local 224 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x24() local
|
| D | vcvt-avx-x24.c | 178 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x24() local 228 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x24() local
|
| D | vcvt-sse41-x24.c | 178 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x24() local 228 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x24() local
|
| D | vcvt-sse2-x32.c | 215 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x32() local 265 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x32() local
|
| D | vcvt-wasmsimd-x32.c | 208 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x32() local 258 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x32() local
|
| D | vcvt-sse41-x32.c | 212 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x32() local 262 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x32() local
|
| D | vcvt-avx-x32.c | 212 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x32() local 262 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x32() local
|
| /external/XNNPACK/src/f32-f16-vcvt/ |
| D | wasmsimd.c.in | 114 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); variable 164 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); variable
|
| /external/XNNPACK/src/math/ |
| D | cvt-f32-f16-neon.c | 40 uint32x4_t vbias_lo = vaddq_u32(vreinterpretq_u32_f32(vabsx_lo), vexp_bias); in xnn_math_f32_f16_cvt__neon() local
|
| D | cvt-f32-f16-sse2.c | 43 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_math_f32_f16_cvt__sse2() local
|
| D | cvt-f32-f16-wasmsimd.c | 42 v128_t vbias_lo = wasm_i32x4_add(vabsx_lo, vexp_bias); in xnn_math_f32_f16_cvt__wasmsimd() local
|
| D | cvt-f32-f16-sse41.c | 43 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_math_f32_f16_cvt__sse41() local
|
| /external/XNNPACK/src/amalgam/ |
| D | sse41.c | 192 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local 242 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local
|
| D | avx.c | 1393 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x24() local 1443 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__avx_x24() local
|
| D | sse2.c | 1121 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local 1171 __m128i vbias_lo = _mm_add_epi32(_mm_castps_si128(vabsx_lo), vexp_bias); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local
|