/external/XNNPACK/src/math/ |
D | exp-avx2-rr2-lut8-p3-perm.c | 60 __m256i veo = _mm256_slli_epi32(_mm256_and_si256(_mm256_castps_si256(vn), vmantissa_mask), 20); in xnn_math_f32_exp__avx2_rr2_lut8_p3_perm() local 61 __m256i ven = _mm256_max_epi32(veo, vmin_exponent); in xnn_math_f32_exp__avx2_rr2_lut8_p3_perm() 63 veo = _mm256_sub_epi32(veo, ven); in xnn_math_f32_exp__avx2_rr2_lut8_p3_perm() 65 const __m256 vso = _mm256_castsi256_ps(_mm256_add_epi32(veo, vdefault_exponent)); in xnn_math_f32_exp__avx2_rr2_lut8_p3_perm()
|
D | exp-avx512f-rr2-lut32-p2-perm2.c | 71 __m512i veo = _mm512_slli_epi32(_mm512_and_si512(_mm512_castps_si512(vn), vmantissa_mask), 18); in xnn_math_f32_exp__avx512f_rr2_lut32_p2_perm2() local 72 __m512i ven = _mm512_max_epi32(veo, vmin_exponent); in xnn_math_f32_exp__avx512f_rr2_lut32_p2_perm2() 74 veo = _mm512_sub_epi32(veo, ven); in xnn_math_f32_exp__avx512f_rr2_lut32_p2_perm2() 76 const __m512 vso = _mm512_castsi512_ps(_mm512_maskz_add_epi32(vinvuf, veo, vdefault_exponent)); in xnn_math_f32_exp__avx512f_rr2_lut32_p2_perm2()
|
D | exp-avx512f-rr2-lut16-p3-perm.c | 66 __m512i veo = _mm512_slli_epi32(_mm512_and_si512(_mm512_castps_si512(vn), vmantissa_mask), 19); in xnn_math_f32_exp__avx512f_rr2_lut16_p3_perm() local 67 __m512i ven = _mm512_max_epi32(veo, vmin_exponent); in xnn_math_f32_exp__avx512f_rr2_lut16_p3_perm() 69 veo = _mm512_sub_epi32(veo, ven); in xnn_math_f32_exp__avx512f_rr2_lut16_p3_perm() 71 const __m512 vso = _mm512_castsi512_ps(_mm512_maskz_add_epi32(vinvuf, veo, vdefault_exponent)); in xnn_math_f32_exp__avx512f_rr2_lut16_p3_perm()
|
D | exp-avx2-rr2-p5.c | 59 __m256i veo = _mm256_slli_epi32(_mm256_castps_si256(vn), 23); in xnn_math_f32_exp__avx2_rr2_p5() local 60 __m256i ven = _mm256_max_epi32(veo, vmin_exponent); in xnn_math_f32_exp__avx2_rr2_p5() 62 veo = _mm256_sub_epi32(veo, ven); in xnn_math_f32_exp__avx2_rr2_p5() 64 const __m256 vso = _mm256_castsi256_ps(_mm256_add_epi32(veo, vdefault_exponent)); in xnn_math_f32_exp__avx2_rr2_p5()
|
D | exp-avx512f-rr2-p5.c | 59 __m512i veo = _mm512_slli_epi32(_mm512_castps_si512(vn), 23); in xnn_math_f32_exp__avx512f_rr2_p5() local 60 __m512i ven = _mm512_max_epi32(veo, vmin_exponent); in xnn_math_f32_exp__avx512f_rr2_p5() 62 veo = _mm512_sub_epi32(veo, ven); in xnn_math_f32_exp__avx512f_rr2_p5() 64 const __m512 vso = _mm512_castsi512_ps(_mm512_add_epi32(veo, vdefault_exponent)); in xnn_math_f32_exp__avx512f_rr2_p5()
|
D | exp-neonfma-rr2-p5.c | 60 int32x4_t veo = vshlq_n_s32(vreinterpretq_s32_f32(vn), 23); in xnn_math_f32_exp__neonfma_rr2_p5() local 61 int32x4_t ven = vmaxq_s32(veo, vmin_exponent); in xnn_math_f32_exp__neonfma_rr2_p5() 63 veo = vsubq_s32(veo, ven); in xnn_math_f32_exp__neonfma_rr2_p5() 65 const float32x4_t vso = vreinterpretq_f32_s32(vaddq_s32(veo, vdefault_exponent)); in xnn_math_f32_exp__neonfma_rr2_p5()
|
D | exp-sse2-rr2-p5.c | 60 __m128i veo = _mm_slli_epi32(_mm_castps_si128(vn), 23); in xnn_math_f32_exp__sse2_rr2_p5() local 61 __m128i ven = _mm_max_epi16(veo, vmin_exponent); in xnn_math_f32_exp__sse2_rr2_p5() 63 veo = _mm_sub_epi32(veo, ven); in xnn_math_f32_exp__sse2_rr2_p5() 65 const __m128 vso = _mm_castsi128_ps(_mm_add_epi32(veo, vdefault_exponent)); in xnn_math_f32_exp__sse2_rr2_p5()
|
D | exp-avx2-rr2-lut8-p4-perm.c | 63 __m256i veo = _mm256_slli_epi32(_mm256_and_si256(_mm256_castps_si256(vn), vmantissa_mask), 20); in xnn_math_f32_exp__avx2_rr2_lut8_p4_perm() local 64 __m256i ven = _mm256_max_epi32(veo, vmin_exponent); in xnn_math_f32_exp__avx2_rr2_lut8_p4_perm() 66 veo = _mm256_sub_epi32(veo, ven); in xnn_math_f32_exp__avx2_rr2_lut8_p4_perm() 68 const __m256 vso = _mm256_castsi256_ps(_mm256_add_epi32(veo, vdefault_exponent)); in xnn_math_f32_exp__avx2_rr2_lut8_p4_perm()
|
D | exp-neonfma-rr2-lut64-p2.c | 62 … int32x4_t veo = vshlq_n_s32(vbicq_s32(vreinterpretq_s32_f32(vn), vmovq_n_s32(INT32_C(0x3F))), 17); in xnn_math_f32_exp__neonfma_rr2_lut64_p2() local 63 int32x4_t ven = vmaxq_s32(veo, vmin_exponent); in xnn_math_f32_exp__neonfma_rr2_lut64_p2() 65 veo = vsubq_s32(veo, ven); in xnn_math_f32_exp__neonfma_rr2_lut64_p2() 78 vl = vreinterpretq_f32_s32(vaddq_s32(vreinterpretq_s32_f32(vl), veo)); in xnn_math_f32_exp__neonfma_rr2_lut64_p2()
|
D | exp-sse2-rr2-lut64-p2.c | 62 __m128i veo = _mm_slli_epi32(_mm_andnot_si128(vindex_mask, _mm_castps_si128(vn)), 17); in xnn_math_f32_exp__sse2_rr2_lut64_p2() local 63 __m128i ven = _mm_max_epi16(veo, vmin_exponent); in xnn_math_f32_exp__sse2_rr2_lut64_p2() 65 veo = _mm_sub_epi32(veo, ven); in xnn_math_f32_exp__sse2_rr2_lut64_p2() 88 …mm_add_epi32(_mm_unpacklo_epi64(_mm_unpacklo_epi32(vl0, vl1), _mm_unpacklo_epi32(vl2, vl3)), veo)); in xnn_math_f32_exp__sse2_rr2_lut64_p2()
|
/external/llvm-project/openmp/libomptarget/cmake/Modules/ |
D | LibomptargetGetDependencies.cmake | 189 veo
|
/external/cldr/tools/java/org/unicode/cldr/util/data/ |
D | iso-639-3_Retirements.tab | 116 …oi], Cruzeño [crz], Ineseño [inz], Obispeño [obi], Purisimeño [puy], and Ventureño [veo] 2009-01-16
|
D | iso-639-3_Name_Index.tab | 7151 veo Ventureño Ventureño
|
D | iso-639-3.tab | 6847 veo I E Ventureño
|
D | language-subtag-registry | 36961 Subtag: veo
|
/external/icu/icu4c/source/samples/ufortune/resources/ |
D | es.txt | 145 "En Argentina veo que el mercado se está expandiendo rápido, casi todo el mundo habla de Internet y…
|
/external/icu/icu4c/source/data/zone/ |
D | ga.txt | 1145 ec{"Antananairíveo"}
|
/external/python/google-api-python-client/samples/prediction/ |
D | language_id.txt | 71 …os he tenido por discreto y prudente en todas vuestras aciones. Pero agora veo que est�is tan lejo… 246 …in acotaciones en las m�rgenes y sin anotaciones en el fin del libro, como veo que est�n otros lib…
|
/external/icu/icu4c/source/data/misc/ |
D | supplementalData.txt | 7041 "veo~p", 24025 "veo",
|
/external/cldr/tools/java/org/unicode/cldr/util/data/languages/ |
D | entityToCode.tsv | 2749 http://www.wikidata.org/entity/Q56712 veo
|