/external/XNNPACK/src/f32-dwconv-spchw/ |
D | 3x3p1-sse.c | 130 __m128 vo = _mm_add_ps(vo4567p0, vo4567p1); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() local 131 vo = _mm_add_ps(vo, vo4567p2); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() 133 vo = _mm_max_ps(vo, vmin); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() 134 vo = _mm_min_ps(vo, vmax); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() 136 _mm_storeu_ps(output, vo); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() 190 __m128 vo = _mm_add_ps(vo4567p0, vo4567p1); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() local 191 vo = _mm_add_ps(vo, vo4567p2); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() 193 vo = _mm_max_ps(vo, vmin); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() 194 vo = _mm_min_ps(vo, vmax); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() 197 _mm_storeu_ps(output, vo); in xnn_f32_dwconv_spchw_ukernel_3x3p1__sse() [all …]
|
D | 3x3s2p1-sse.c | 106 __m128 vo = _mm_add_ps(vo8ACEp0, vo8ACEp1); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() local 107 vo = _mm_add_ps(vo, vo8ACEp2); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() 109 vo = _mm_max_ps(vo, vmin); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() 110 vo = _mm_min_ps(vo, vmax); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() 112 _mm_storeu_ps(output, vo); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() 155 __m128 vo = _mm_add_ps(vo8ACEp0, vo8ACEp1); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() local 156 vo = _mm_add_ps(vo, vo8ACEp2); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() 158 vo = _mm_max_ps(vo, vmin); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() 159 vo = _mm_min_ps(vo, vmax); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() 162 _mm_storeu_ps(output, vo); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__sse() [all …]
|
D | 3x3s2p1-neonfma.c | 94 float32x4_t vo = vaddq_f32(vo468Ap0, vo468Ap1); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() local 95 vo = vaddq_f32(vo, vo468Ap2); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 97 vo = vmaxq_f32(vo, vmin); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 98 vo = vminq_f32(vo, vmax); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 100 vst1q_f32(output, vo); output = (float*) ((uintptr_t) output + output_tuple_stride); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 143 float32x4_t vo = vaddq_f32(vo468Ap0, vo468Ap1); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() local 144 vo = vaddq_f32(vo, vo468Ap2); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 146 vo = vmaxq_f32(vo, vmin); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 147 vo = vminq_f32(vo, vmax); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 151 vst1q_f32(output, vo); in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() [all …]
|
D | 3x3p1-neonfma.c | 307 float32x4_t vo = vaddq_f32(vo4567p0, vo4567p1); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() local 308 vo = vaddq_f32(vo, vo4567p2); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 310 vo = vmaxq_f32(vo, vmin); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 311 vo = vminq_f32(vo, vmax); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 313 vst1q_f32(output0, vo); output0 = (float*) ((uintptr_t) output0 + output_tuple_stride); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 346 float32x4_t vo = vaddq_f32(vo4567p0, vo4567p1); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() local 347 vo = vaddq_f32(vo, vo4567p2); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 349 vo = vmaxq_f32(vo, vmin); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 350 vo = vminq_f32(vo, vmax); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 353 vst1q_f32(output0, vo); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/ARM/ |
D | unaligned_load_store_vector.ll | 12 %vo = bitcast i8* %po to <8 x i8>* 16 store <8 x i8> %v1, <8 x i8>* %vo, align 1 30 %vo = bitcast i8* %po to <4 x i16>* 34 store <4 x i16> %v1, <4 x i16>* %vo, align 1 48 %vo = bitcast i8* %po to <2 x i32>* 52 store <2 x i32> %v1, <2 x i32>* %vo, align 1 66 %vo = bitcast i8* %po to <2 x float>* 70 store <2 x float> %v1, <2 x float>* %vo, align 1 84 %vo = bitcast i8* %po to <16 x i8>* 88 store <16 x i8> %v1, <16 x i8>* %vo, align 1 [all …]
|
/external/llvm/test/CodeGen/ARM/ |
D | unaligned_load_store_vector.ll | 12 %vo = bitcast i8* %po to <8 x i8>* 16 store <8 x i8> %v1, <8 x i8>* %vo, align 1 30 %vo = bitcast i8* %po to <4 x i16>* 34 store <4 x i16> %v1, <4 x i16>* %vo, align 1 48 %vo = bitcast i8* %po to <2 x i32>* 52 store <2 x i32> %v1, <2 x i32>* %vo, align 1 66 %vo = bitcast i8* %po to <2 x float>* 70 store <2 x float> %v1, <2 x float>* %vo, align 1 84 %vo = bitcast i8* %po to <16 x i8>* 88 store <16 x i8> %v1, <16 x i8>* %vo, align 1 [all …]
|
/external/llvm/test/CodeGen/SystemZ/ |
D | vec-or-01.ll | 8 ; CHECK: vo %v24, %v26, %v28 17 ; CHECK: vo %v24, %v26, %v28 26 ; CHECK: vo %v24, %v26, %v28 35 ; CHECK: vo %v24, %v26, %v28
|
D | vec-cmp-06.ll | 20 ; CHECK: vo %v24, [[REG1]], [[REG2]] 143 ; CHECK: vo %v24, [[REG1]], [[REG2]] 181 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]] 243 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]] 317 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]] 331 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/SystemZ/ |
D | vec-or-01.ll | 8 ; CHECK: vo %v24, %v26, %v28 17 ; CHECK: vo %v24, %v26, %v28 26 ; CHECK: vo %v24, %v26, %v28 35 ; CHECK: vo %v24, %v26, %v28
|
D | vec-cmp-06.ll | 20 ; CHECK: vo %v24, [[REG1]], [[REG2]] 143 ; CHECK: vo %v24, [[REG1]], [[REG2]] 181 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]] 243 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]] 317 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]] 331 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]]
|
D | vec-cmp-07.ll | 20 ; CHECK: vo %v24, [[REG1]], [[REG2]] 143 ; CHECK: vo %v24, [[REG1]], [[REG2]] 181 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]] 243 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]] 317 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]] 331 ; CHECK: vo [[REG:%v[0-9]+]], [[REG1]], [[REG2]]
|
/external/XNNPACK/src/f32-bilinear/ |
D | scalar.c.in | 63 const float vo${ABC[C]} = vt${ABC[C]} + vd${ABC[C]} * valphav; 66 output[${C}] = vo${ABC[C]}; 83 const float vo = vt + vd * valphav; 85 *output++ = vo; 102 const float vo = vt + vd * valphav; 104 *output++ = vo;
|
D | neon.c.in | 79 const float32x4_t vo${ABC[C:C+4]} = vfmaq_f32(vt${ABC[C:C+4]}, vd${ABC[C:C+4]}, valphav); 82 … const float32x4_t vo${ABC[C:C+4]} = vfmaq_lane_f32(vt${ABC[C:C+4]}, vd${ABC[C:C+4]}, valphahv, 1); 86 … const float32x4_t vo${ABC[C:C+4]} = vmlaq_lane_f32(vt${ABC[C:C+4]}, vd${ABC[C:C+4]}, valphahv, 1); 89 vst1q_f32(output, vo${ABC[C:C+4]}); output += 4;
|
D | psimd.c.in | 69 const psimd_f32 vo${ABC[C:C+4]} = psimd_qfma_f32(vt${ABC[C:C+4]}, vd${ABC[C:C+4]}, valphav); 71 psimd_store_f32(output, vo${ABC[0:4]}); 73 psimd_store_f32(output + ${C}, vo${ABC[C:C+4]});
|
D | sse.c.in | 71 … const __m128 vo${ABC[C:C+4]} = _mm_add_ps(vt${ABC[C:C+4]}, _mm_mul_ps(vd${ABC[C:C+4]}, valphav)); 73 _mm_storeu_ps(output, vo${ABC[0:4]}); 75 _mm_storeu_ps(output + ${C}, vo${ABC[C:C+4]});
|
/external/cldr/common/testData/transforms/ |
D | hy-fonipa-t-hy.txt | 24 որևէ voɾɛvɛ 29 որպես voɾpɛs 38 որոնք voɾonkʰ 58 որը voɾə 80 որևիցէ voɾɛvit͡sʰɛ 81 որի voɾi 82 որ voɾ 117 որպեսզի voɾpɛszi 118 որով voɾov 201 տրվող tɾvoɣ [all …]
|
/external/XNNPACK/src/f32-bilinear/gen/ |
D | scalar-c1.c | 54 const float vo = vt + vd * valphav; in xnn_f32_bilinear_ukernel__scalar_c1() local 56 *output++ = vo; in xnn_f32_bilinear_ukernel__scalar_c1()
|
D | scalar-c2.c | 88 const float vo = vt + vd * valphav; in xnn_f32_bilinear_ukernel__scalar_c2() local 90 *output++ = vo; in xnn_f32_bilinear_ukernel__scalar_c2()
|
D | scalar-c4.c | 110 const float vo = vt + vd * valphav; in xnn_f32_bilinear_ukernel__scalar_c4() local 112 *output++ = vo; in xnn_f32_bilinear_ukernel__scalar_c4()
|
/external/XNNPACK/src/f32-maxpool/ |
D | 9p8x-sse-c4.c | 210 const __m128 vo = _mm_loadu_ps(o); in xnn_f32_maxpool_ukernel_9p8x__sse_c4() local 212 const __m128 vmax01 = _mm_max_ps(_mm_max_ps(vi0, vi1), vo); in xnn_f32_maxpool_ukernel_9p8x__sse_c4() 234 const __m128 vo = _mm_loadu_ps(o); in xnn_f32_maxpool_ukernel_9p8x__sse_c4() local 236 const __m128 vmax01 = _mm_max_ps(_mm_max_ps(vi0, vi1), vo); in xnn_f32_maxpool_ukernel_9p8x__sse_c4()
|
D | 9p8x-psimd-c4.c | 210 const psimd_f32 vo = psimd_load_f32(o); in xnn_f32_maxpool_ukernel_9p8x__psimd_c4() local 212 const psimd_f32 vmax01 = psimd_max_f32(psimd_max_f32(vi0, vi1), vo); in xnn_f32_maxpool_ukernel_9p8x__psimd_c4() 234 const psimd_f32 vo = psimd_load_f32(o); in xnn_f32_maxpool_ukernel_9p8x__psimd_c4() local 236 const psimd_f32 vmax01 = psimd_max_f32(psimd_max_f32(vi0, vi1), vo); in xnn_f32_maxpool_ukernel_9p8x__psimd_c4()
|
/external/XNNPACK/src/u8-maxpool/ |
D | 9p8x-neon-c16.c | 193 const uint8x16_t vo = vld1q_u8(o); in xnn_u8_maxpool_ukernel_9p8x__neon_c16() local 195 const uint8x16_t vmax01 = vmaxq_u8(vmaxq_u8(vi0, vi1), vo); in xnn_u8_maxpool_ukernel_9p8x__neon_c16() 216 const uint8x16_t vo = vld1q_u8(o); in xnn_u8_maxpool_ukernel_9p8x__neon_c16() local 218 const uint8x16_t vmax01 = vmaxq_u8(vmaxq_u8(vi0, vi1), vo); in xnn_u8_maxpool_ukernel_9p8x__neon_c16()
|
D | 9p8x-sse2-c16.c | 197 const __m128i vo = _mm_loadu_si128((const __m128i*) o); in xnn_u8_maxpool_ukernel_9p8x__sse2_c16() local 199 const __m128i vmax01 = _mm_max_epu8(_mm_max_epu8(vi0, vi1), vo); in xnn_u8_maxpool_ukernel_9p8x__sse2_c16() 221 const __m128i vo = _mm_loadu_si128((const __m128i*) o); in xnn_u8_maxpool_ukernel_9p8x__sse2_c16() local 223 const __m128i vmax01 = _mm_max_epu8(_mm_max_epu8(vi0, vi1), vo); in xnn_u8_maxpool_ukernel_9p8x__sse2_c16()
|
/external/error_prone/checkerframework/ |
D | dataflow-2.5.3.jar.asc | 9 +k+ac7utiirqp1F2/vo/0QwEygdDI+q6RFk9vS7YclUJoa7InaQdxqalrtpBNiU=
|
/external/icu/icu4c/source/data/translit/ |
D | hy_hy_FONIPA.txt | 47 $wordBoundary {ո} → vo;
|