| /external/XNNPACK/src/f32-f16-vcvt/gen/ |
| D | vcvt-sse2-x8.c | 75 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x8() local 125 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x8() local
|
| D | vcvt-avx-x8.c | 73 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x8() local 123 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x8() local
|
| D | vcvt-sse41-x8.c | 73 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local 123 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local
|
| D | vcvt-wasmsimd-x8.c | 71 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x8() local 121 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x8() local
|
| D | vcvt-avx-x16.c | 166 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x16() local 216 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x16() local
|
| D | vcvt-wasmsimd-x16.c | 162 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x16() local 212 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x16() local
|
| D | vcvt-sse41-x16.c | 166 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x16() local 216 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x16() local
|
| D | vcvt-sse2-x16.c | 169 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local 219 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local
|
| D | vcvt-sse2-x24.c | 204 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x24() local 254 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x24() local
|
| D | vcvt-wasmsimd-x24.c | 196 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x24() local 246 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x24() local
|
| D | vcvt-avx-x24.c | 200 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x24() local 250 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x24() local
|
| D | vcvt-sse41-x24.c | 200 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x24() local 250 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x24() local
|
| D | vcvt-sse2-x32.c | 239 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x32() local 289 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x32() local
|
| D | vcvt-wasmsimd-x32.c | 230 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x32() local 280 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); in xnn_f32_f16_vcvt_ukernel__wasmsimd_x32() local
|
| D | vcvt-sse41-x32.c | 234 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x32() local 284 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x32() local
|
| D | vcvt-avx-x32.c | 234 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x32() local 284 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x32() local
|
| /external/XNNPACK/src/f32-f16-vcvt/ |
| D | wasmsimd.c.in | 136 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); variable 186 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); variable
|
| /external/XNNPACK/src/math/ |
| D | cvt-f32-f16-sse2.c | 67 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_math_f32_f16_cvt__sse2() local
|
| D | cvt-f32-f16-wasmsimd.c | 64 const v128_t vmantw_lo = wasm_v128_and(vf_lo, vmanth_mask); in xnn_math_f32_f16_cvt__wasmsimd() local
|
| D | cvt-f32-f16-sse41.c | 65 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_math_f32_f16_cvt__sse41() local
|
| /external/XNNPACK/src/amalgam/ |
| D | sse41.c | 214 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local 264 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse41_x8() local
|
| D | avx.c | 1415 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x24() local 1465 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__avx_x24() local
|
| D | sse2.c | 1145 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local 1195 const __m128i vmantw_lo = _mm_and_si128(_mm_castps_si128(vf_lo), vmanth_mask); in xnn_f32_f16_vcvt_ukernel__sse2_x16() local
|