Searched refs:vs1_lo (Results 1 – 10 of 10) sorted by relevance
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-avx-rr2-p6-x16.c | 58 …const __m128 vs1_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn1)… in xnn_f32_velu_ukernel__avx_rr2_p6_x16() local 65 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x16()
|
D | velu-avx-rr2-p6-x24.c | 61 …const __m128 vs1_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn1)… in xnn_f32_velu_ukernel__avx_rr2_p6_x24() local 71 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x24()
|
D | velu-avx-rr2-p6-x32.c | 64 …const __m128 vs1_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn1)… in xnn_f32_velu_ukernel__avx_rr2_p6_x32() local 77 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x32()
|
D | velu-avx-rr2-p6-x40.c | 67 …const __m128 vs1_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn1)… in xnn_f32_velu_ukernel__avx_rr2_p6_x40() local 83 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x40()
|
D | velu-avx-rr2-p6-x48.c | 70 …const __m128 vs1_lo = _mm_castsi128_ps(_mm_slli_epi32(_mm_castps_si128(_mm256_castps256_ps128(vn1)… in xnn_f32_velu_ukernel__avx_rr2_p6_x48() local 89 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_p6_x48()
|
D | velu-avx-rr2-lut16-p3-x16.c | 123 const __m128 vs1_lo = _mm_castsi128_ps(_mm_add_epi32(vl1_lo, ven1_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x16() local 132 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x16()
|
D | velu-avx-rr2-lut16-p3-x24.c | 155 const __m128 vs1_lo = _mm_castsi128_ps(_mm_add_epi32(vl1_lo, ven1_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24() local 170 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x24()
|
D | velu-avx-rr2-lut16-p3-x32.c | 187 const __m128 vs1_lo = _mm_castsi128_ps(_mm_add_epi32(vl1_lo, ven1_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32() local 208 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x32()
|
D | velu-avx-rr2-lut16-p3-x40.c | 219 const __m128 vs1_lo = _mm_castsi128_ps(_mm_add_epi32(vl1_lo, ven1_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40() local 246 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x40()
|
D | velu-avx-rr2-lut16-p3-x48.c | 251 const __m128 vs1_lo = _mm_castsi128_ps(_mm_add_epi32(vl1_lo, ven1_lo)); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48() local 284 __m256 vs1 = _mm256_insertf128_ps(_mm256_castps128_ps256(vs1_lo), vs1_hi, 1); in xnn_f32_velu_ukernel__avx_rr2_lut16_p3_x48()
|