/external/llvm-project/clang/test/Sema/ |
D | ext_vector_conversions.c | 14 short4 vs4; in test() local
|
D | vecshift.c | 47 vector_short4 vs4; variable
|
/external/clang/test/Sema/ |
D | ext_vector_conversions.c | 14 short4 vs4; in test() local
|
/external/llvm-project/clang/test/CodeGen/ |
D | vecshift.c | 45 vector_short4 vs4; variable
|
/external/XNNPACK/src/f32-velu/gen/ |
D | velu-scalar-rr2-p6-x5.c | 71 float vs4 = fp32_from_bits(fp32_to_bits(vn4) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x5() local
|
D | velu-wasm-rr2-p6-x5.c | 71 float vs4 = fp32_from_bits(fp32_to_bits(vn4) << 23); in xnn_f32_velu_ukernel__wasm_rr2_p6_x5() local
|
D | velu-scalar-rr2-p6-x6.c | 74 float vs4 = fp32_from_bits(fp32_to_bits(vn4) << 23); in xnn_f32_velu_ukernel__scalar_rr2_p6_x6() local
|
D | velu-avx512f-rr1-p6-x80.c | 70 __m512 vs4 = _mm512_castsi512_ps(_mm512_slli_epi32(_mm512_castps_si512(vn4), 23)); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x80() local
|
D | velu-wasm-rr2-p6-x6.c | 74 float vs4 = fp32_from_bits(fp32_to_bits(vn4) << 23); in xnn_f32_velu_ukernel__wasm_rr2_p6_x6() local
|
D | velu-avx2-rr1-p6-x40.c | 70 __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn4), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x40() local
|
D | velu-scalar-rr2-lut16-p3-x5.c | 88 float vs4 = fp32_from_bits(xnn_table_exp2minus_k_over_16[vidx4] + ven4); in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x5() local
|
D | velu-wasm-rr2-lut16-p3-x5.c | 88 float vs4 = fp32_from_bits(xnn_table_exp2minus_k_over_16[vidx4] + ven4); in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x5() local
|
D | velu-avx512f-rr1-p6-x96.c | 73 __m512 vs4 = _mm512_castsi512_ps(_mm512_slli_epi32(_mm512_castps_si512(vn4), 23)); in xnn_f32_velu_ukernel__avx512f_rr1_p6_x96() local
|
D | velu-avx512f-rr1-lut16-p3-perm-x80.c | 81 __m512 vs4 = _mm512_castsi512_ps(_mm512_add_epi32(vl4, ven4)); in xnn_f32_velu_ukernel__avx512f_rr1_lut16_p3_perm_x80() local
|
D | velu-scalar-rr2-lut16-p3-x6.c | 94 float vs4 = fp32_from_bits(xnn_table_exp2minus_k_over_16[vidx4] + ven4); in xnn_f32_velu_ukernel__scalar_rr2_lut16_p3_x6() local
|
D | velu-avx2-rr1-p6-x48.c | 73 __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn4), 23)); in xnn_f32_velu_ukernel__avx2_rr1_p6_x48() local
|
D | velu-wasm-rr2-lut16-p3-x6.c | 94 float vs4 = fp32_from_bits(xnn_table_exp2minus_k_over_16[vidx4] + ven4); in xnn_f32_velu_ukernel__wasm_rr2_lut16_p3_x6() local
|
/external/XNNPACK/src/f32-vscaleextexp/gen/ |
D | avx2-p5-x40.c | 144 …const __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve4, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x40() local
|
D | avx2-p5-x48.c | 156 …const __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve4, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x48() local
|
D | avx2-p5-x56.c | 168 …const __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(_mm256_add_ps(ve4, vm… in xnn_f32_vscaleextexp_ukernel__avx2_p5_x56() local
|
/external/XNNPACK/src/f32-vscaleexpminusmax/gen/ |
D | avx2-p5-x40.c | 74 const __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn4), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x40() local
|
D | avx2-p5-x48.c | 77 const __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn4), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x48() local
|
D | avx2-p5-x56.c | 80 const __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn4), 23)); in xnn_f32_vscaleexpminusmax_ukernel__avx2_p5_x56() local
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | avx2-rr1-p5-div-x40.c | 64 const __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn4), 23)); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_div_x40() local
|
D | avx2-rr1-p5-nr2fma-x40.c | 64 const __m256 vs4 = _mm256_castsi256_ps(_mm256_slli_epi32(_mm256_castps_si256(vn4), 23)); in xnn_f32_sigmoid_ukernel__avx2_rr1_p5_nr2fma_x40() local
|