/external/XNNPACK/src/f32-gavgpool-spchw/ |
D | neon-x4.c | 60 vi0 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi0))); in xnn_f32_gavgpool_spchw_ukernel__neon_x4() 61 vi1 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi1))); in xnn_f32_gavgpool_spchw_ukernel__neon_x4() 62 vi2 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi2))); in xnn_f32_gavgpool_spchw_ukernel__neon_x4() 63 vi3 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi3))); in xnn_f32_gavgpool_spchw_ukernel__neon_x4() 110 vi0 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi0))); in xnn_f32_gavgpool_spchw_ukernel__neon_x4()
|
/external/XNNPACK/src/f32-dwconv-spchw/ |
D | 3x3s2p1-neonfma.c | 115 …const float32x4_t vi0x468A = vreinterpretq_f32_u32(vandq_u32(vmask_even, vreinterpretq_u32_f32(vuz… in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 116 …const float32x4_t vi0x579B = vreinterpretq_f32_u32(vandq_u32(vmask_odd, vreinterpretq_u32_f32(vuz… in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 117 …const float32x4_t vi1x468A = vreinterpretq_f32_u32(vandq_u32(vmask_even, vreinterpretq_u32_f32(vuz… in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 118 …const float32x4_t vi1x579B = vreinterpretq_f32_u32(vandq_u32(vmask_odd, vreinterpretq_u32_f32(vuz… in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 119 …const float32x4_t vi2x468A = vreinterpretq_f32_u32(vandq_u32(vmask_even, vreinterpretq_u32_f32(vuz… in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma() 120 …const float32x4_t vi2x579B = vreinterpretq_f32_u32(vandq_u32(vmask_odd, vreinterpretq_u32_f32(vuz… in xnn_f32_dwconv_spchw_ukernel_3x3s2p1__neonfma()
|
D | 5x5s2p2-neonfma.c | 256 vi0x468A = vreinterpretq_f32_u32(vandq_u32(vmask_even, vreinterpretq_u32_f32(vi0x468A))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma() 257 vi1x468A = vreinterpretq_f32_u32(vandq_u32(vmask_even, vreinterpretq_u32_f32(vi1x468A))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma() 258 vi2x468A = vreinterpretq_f32_u32(vandq_u32(vmask_even, vreinterpretq_u32_f32(vi2x468A))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma() 259 vi3x468A = vreinterpretq_f32_u32(vandq_u32(vmask_even, vreinterpretq_u32_f32(vi3x468A))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma() 260 vi4x468A = vreinterpretq_f32_u32(vandq_u32(vmask_even, vreinterpretq_u32_f32(vi4x468A))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma() 262 vi0x579B = vreinterpretq_f32_u32(vandq_u32(vmask_odd, vreinterpretq_u32_f32(vi0x579B))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma() 263 vi1x579B = vreinterpretq_f32_u32(vandq_u32(vmask_odd, vreinterpretq_u32_f32(vi1x579B))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma() 264 vi2x579B = vreinterpretq_f32_u32(vandq_u32(vmask_odd, vreinterpretq_u32_f32(vi2x579B))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma() 265 vi3x579B = vreinterpretq_f32_u32(vandq_u32(vmask_odd, vreinterpretq_u32_f32(vi3x579B))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma() 266 vi4x579B = vreinterpretq_f32_u32(vandq_u32(vmask_odd, vreinterpretq_u32_f32(vi4x579B))); in xnn_f32_dwconv_spchw_ukernel_5x5s2p2__neonfma()
|
D | 5x5p2-neonfma.c | 265 vi0x89AB = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi0x89AB))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() 266 vi1x89AB = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi1x89AB))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() 267 vi2x89AB = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi2x89AB))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() 268 vi3x89AB = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi3x89AB))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() 269 vi4x89AB = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi4x89AB))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() 270 vi5x89AB = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi5x89AB))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() 271 vi6x89AB = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi6x89AB))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() 449 vi0x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi0x4567))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() 450 vi1x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi1x4567))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() 451 vi2x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi2x4567))); in xnn_f32_dwconv_spchw_ukernel_5x5p2__neonfma() [all …]
|
D | 3x3p1-neonfma.c | 159 vi0x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi0x4567))); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 160 vi1x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi1x4567))); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 161 vi2x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi2x4567))); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 162 vi3x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi3x4567))); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 163 vi4x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi4x4567))); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 321 vi0x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi0x4567))); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 322 vi1x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi1x4567))); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma() 323 vi2x4567 = vreinterpretq_f32_u32(vandq_u32(vmask, vreinterpretq_u32_f32(vi2x4567))); in xnn_f32_dwconv_spchw_ukernel_3x3p1__neonfma()
|
/external/webrtc/webrtc/modules/audio_processing/aec/ |
D | aec_core_neon.c | 113 x = vreinterpretq_f32_u32(vandq_u32(vmvnq_u32(div_by_zero), in vsqrtq_f32() 157 uint32x4_t ef_re_u32 = vandq_u32(vmvnq_u32(bigger), in ScaleErrorSignalNEON() 159 uint32x4_t ef_im_u32 = vandq_u32(vmvnq_u32(bigger), in ScaleErrorSignalNEON() 161 ef_re_if = vandq_u32(bigger, ef_re_if); in ScaleErrorSignalNEON() 162 ef_im_if = vandq_u32(bigger, ef_im_if); in ScaleErrorSignalNEON() 295 const uint32x4_t two_n = vandq_u32(vreinterpretq_u32_f32(a), in vpowq_f32() 305 const uint32x4_t mantissa = vandq_u32(vreinterpretq_u32_f32(a), in vpowq_f32() 403 const uint32x4_t vec_if0 = vandq_u32(vmvnq_u32(bigger), in OverdriveAndSuppressNEON() 408 vandq_u32(bigger, vreinterpretq_u32_f32(vec_one_weightCurve_add)); in OverdriveAndSuppressNEON()
|
/external/libaom/libaom/av1/common/arm/ |
D | selfguided_neon.c | 64 p0 = vandq_u32(p0, q0); in calc_ab_fast_internal_common() 65 p1 = vandq_u32(p1, q1); in calc_ab_fast_internal_common() 66 p2 = vandq_u32(p2, q2); in calc_ab_fast_internal_common() 67 p3 = vandq_u32(p3, q3); in calc_ab_fast_internal_common() 175 p0 = vandq_u32(p0, q0); in calc_ab_internal_common() 176 p1 = vandq_u32(p1, q1); in calc_ab_internal_common() 177 p2 = vandq_u32(p2, q2); in calc_ab_internal_common() 178 p3 = vandq_u32(p3, q3); in calc_ab_internal_common() 179 p4 = vandq_u32(p4, q4); in calc_ab_internal_common() 180 p5 = vandq_u32(p5, q5); in calc_ab_internal_common() [all …]
|
/external/eigen/Eigen/src/Core/arch/NEON/ |
D | MathFunctions.h | 51 mask = vandq_u32(mask, vreinterpretq_u32_f32(p4f_1));
|
D | Complex.h | 106 …return Packet2cf(vreinterpretq_f32_u32(vandq_u32(vreinterpretq_u32_f32(a.v),vreinterpretq_u32_f32(…
|
D | PacketMath.h | 224 return vreinterpretq_f32_u32(vandq_u32(vreinterpretq_u32_f32(a),vreinterpretq_u32_f32(b)));
|
/external/webp/src/dsp/ |
D | lossless_enc_neon.c | 119 const uint32x4_t F = vandq_u32(vreinterpretq_u32_s8(E), mask_rb); in TransformColor_NEON()
|
D | lossless_neon.c | 575 const uint32x4_t a0g0 = vandq_u32(vreinterpretq_u32_u8(in), mask_ag); in TransformColorInverse_NEON()
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-3v.c | 102 return vandq_u32(a, b); in test_vandq_u32()
|
D | arm_neon_intrinsics.c | 986 return vandq_u32(a, b); in test_vandq_u32()
|
/external/skia/include/private/ |
D | SkNx_neon.h | 22 return vsubq_f32(roundtrip, (float32x4_t)vandq_u32(too_big, (uint32x4_t)vdupq_n_f32(1))); in emulate_vrndmq_f32() 603 AI SkNx operator & (const SkNx& o) const { return vandq_u32(fVec, o.fVec); }
|
/external/skqp/include/private/ |
D | SkNx_neon.h | 22 return vsubq_f32(roundtrip, (float32x4_t)vandq_u32(too_big, (uint32x4_t)vdupq_n_f32(1))); in emulate_vrndmq_f32() 603 AI SkNx operator & (const SkNx& o) const { return vandq_u32(fVec, o.fVec); }
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | cdef_neon.cc | 353 uint32x4_t c = vandq_u32(mask, Square(vget_low_u16(a))); in CostOdd()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 2123 _NEON2SSESTORAGE uint32x4_t vandq_u32(uint32x4_t a, uint32x4_t b); // VAND q0,q0,q0 15235 _NEON2SSESTORAGE uint32x4_t vandq_u32(uint32x4_t a, uint32x4_t b); // VAND q0,q0,q0 15236 #define vandq_u32 _mm_and_si128 macro
|