Searched refs:vs0_lo (Results 1 – 10 of 10) sorted by relevance
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-avx-rr2-p6-x16.c | 55 …const __m128 vs0_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn0)… in xnn_f32_velu_ukernel__avx_rr2_p6_x16() local 63 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x16()
|
D | velu-avx-rr2-p6-x24.c | 58 …const __m128 vs0_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn0)… in xnn_f32_velu_ukernel__avx_rr2_p6_x24() local 69 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x24()
|
D | velu-avx-rr2-p6-x32.c | 61 …const __m128 vs0_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn0)… in xnn_f32_velu_ukernel__avx_rr2_p6_x32() local 75 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x32()
|
D | velu-avx-rr2-p6-x40.c | 64 …const __m128 vs0_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn0)… in xnn_f32_velu_ukernel__avx_rr2_p6_x40() local 81 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x40()
|
D | velu-avx-rr2-p6-x48.c | 67 …const __m128 vs0_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn0)… in xnn_f32_velu_ukernel__avx_rr2_p6_x48() local 87 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x48()
|
D | velu-avx-rr2-lut16-p3-x16.c | 118 const __m128 vs0_lo = _mm_castsi128_ps(_mm_add_epi32(vl0_lo, ven0_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x16() local 130 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x16()
|
D | velu-avx-rr2-lut16-p3-x24.c | 150 const __m128 vs0_lo = _mm_castsi128_ps(_mm_add_epi32(vl0_lo, ven0_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24() local 168 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
|
D | velu-avx-rr2-lut16-p3-x32.c | 182 const __m128 vs0_lo = _mm_castsi128_ps(_mm_add_epi32(vl0_lo, ven0_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32() local 206 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
|
D | velu-avx-rr2-lut16-p3-x40.c | 214 const __m128 vs0_lo = _mm_castsi128_ps(_mm_add_epi32(vl0_lo, ven0_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40() local 244 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
|
D | velu-avx-rr2-lut16-p3-x48.c | 246 const __m128 vs0_lo = _mm_castsi128_ps(_mm_add_epi32(vl0_lo, ven0_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48() local 282 __m256 vs0 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs0_lo), vs0_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
|