/external/XNNPACK/src/f32-argmaxpool/ |
D | 9p8x-sse2-c4.c | 78 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 82 const __m128i vm2 = _mm_castps_si128(_mm_cmpgt_ps(vi2, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 86 const __m128i vm3 = _mm_castps_si128(_mm_cmpgt_ps(vi3, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 90 const __m128i vm4 = _mm_castps_si128(_mm_cmpgt_ps(vi4, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 94 const __m128i vm5 = _mm_castps_si128(_mm_cmpgt_ps(vi5, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 98 const __m128i vm6 = _mm_castps_si128(_mm_cmpgt_ps(vi6, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 102 const __m128i vm7 = _mm_castps_si128(_mm_cmpgt_ps(vi7, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 106 const __m128i vm8 = _mm_castps_si128(_mm_cmpgt_ps(vi8, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 163 const __m128i vm0 = _mm_castps_si128(_mm_cmpgt_ps(vi0, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() 167 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9p8x__sse2_c4() [all …]
|
D | 9x-sse2-c4.c | 97 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 101 const __m128i vm2 = _mm_castps_si128(_mm_cmpgt_ps(vi2, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 105 const __m128i vm3 = _mm_castps_si128(_mm_cmpgt_ps(vi3, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 109 const __m128i vm4 = _mm_castps_si128(_mm_cmpgt_ps(vi4, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 113 const __m128i vm5 = _mm_castps_si128(_mm_cmpgt_ps(vi5, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 117 const __m128i vm6 = _mm_castps_si128(_mm_cmpgt_ps(vi6, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 121 const __m128i vm7 = _mm_castps_si128(_mm_cmpgt_ps(vi7, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 125 const __m128i vm8 = _mm_castps_si128(_mm_cmpgt_ps(vi8, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 148 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() 152 const __m128i vm2 = _mm_castps_si128(_mm_cmpgt_ps(vi2, vmax)); in xnn_f32_argmaxpool_ukernel_9x__sse2_c4() [all …]
|
D | 4x-sse2-c4.c | 62 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 66 const __m128i vm2 = _mm_castps_si128(_mm_cmpgt_ps(vi2, vmax)); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 70 const __m128i vm3 = _mm_castps_si128(_mm_cmpgt_ps(vi3, vmax)); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 88 const __m128i vm1 = _mm_castps_si128(_mm_cmpgt_ps(vi1, vmax)); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 92 const __m128i vm2 = _mm_castps_si128(_mm_cmpgt_ps(vi2, vmax)); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4() 96 const __m128i vm3 = _mm_castps_si128(_mm_cmpgt_ps(vi3, vmax)); in xnn_f32_argmaxpool_ukernel_4x__sse2_c4()
|
/external/XNNPACK/src/f32-vrnd/gen/ |
D | vrndd-sse2-x8.c | 47 const __m128 vy0123 = _mm_sub_ps(vrndx0123, _mm_and_ps(_mm_cmpgt_ps(vrndx0123, vx0123), vone)); in xnn_f32_vrndd_ukernel__sse2_x8() 48 const __m128 vy4567 = _mm_sub_ps(vrndx4567, _mm_and_ps(_mm_cmpgt_ps(vrndx4567, vx4567), vone)); in xnn_f32_vrndd_ukernel__sse2_x8() 62 const __m128 vy = _mm_sub_ps(vrndx, _mm_and_ps(_mm_cmpgt_ps(vrndx, vx), vone)); in xnn_f32_vrndd_ukernel__sse2_x8() 73 __m128 vy = _mm_sub_ps(vrndx, _mm_and_ps(_mm_cmpgt_ps(vrndx, vx), vone)); in xnn_f32_vrndd_ukernel__sse2_x8()
|
D | vrndd-sse2-x4.c | 42 const __m128 vy0123 = _mm_sub_ps(vrndx0123, _mm_and_ps(_mm_cmpgt_ps(vrndx0123, vx0123), vone)); in xnn_f32_vrndd_ukernel__sse2_x4() 53 __m128 vy = _mm_sub_ps(vrndx, _mm_and_ps(_mm_cmpgt_ps(vrndx, vx), vone)); in xnn_f32_vrndd_ukernel__sse2_x4()
|
/external/XNNPACK/src/f32-vrnd/ |
D | vrndd-sse2.c.in | 48 …const __m128 vy${ABC[N:N+4]} = _mm_sub_ps(vrndx${ABC[N:N+4]}, _mm_and_ps(_mm_cmpgt_ps(vrndx${ABC[N… 64 const __m128 vy = _mm_sub_ps(vrndx, _mm_and_ps(_mm_cmpgt_ps(vrndx, vx), vone)); 75 __m128 vy = _mm_sub_ps(vrndx, _mm_and_ps(_mm_cmpgt_ps(vrndx, vx), vone));
|
/external/XNNPACK/src/math/ |
D | roundd-sse2-cvt.c | 52 const __m128 vy = _mm_sub_ps(vrndx, _mm_and_ps(_mm_cmpgt_ps(vrndx, vx), vone)); in xnn_math_f32_roundd__sse2_cvt()
|
D | roundd-sse-addsub.c | 57 const __m128 vy = _mm_sub_ps(vrndx, _mm_and_ps(_mm_cmpgt_ps(vrndx, vx), vone)); in xnn_math_f32_roundd__sse_addsub()
|
D | roundz-sse-addsub.c | 52 const __m128 vadjustment = _mm_and_ps(vone, _mm_cmpgt_ps(vrndabsx, vabsx)); in xnn_math_f32_roundz__sse_addsub()
|
D | exp-sse2-rr2-p5.c | 93 const __m128 vm = _mm_cmpgt_ps(vx, vinf_cutoff); in xnn_math_f32_exp__sse2_rr2_p5()
|
D | exp-sse2-rr2-lut64-p2.c | 114 const __m128 vm = _mm_cmpgt_ps(vx, vinf_cutoff); in xnn_math_f32_exp__sse2_rr2_lut64_p2()
|
/external/llvm-project/clang/test/CodeGen/X86/ |
D | sse-builtins-constrained-cmp.c | 30 return _mm_cmpgt_ps(__a, __b); in test_mm_cmpgt_ps()
|
D | sse-builtins.c | 74 return _mm_cmpgt_ps(__a, __b); in test_mm_cmpgt_ps()
|
/external/oboe/samples/RhythmGame/third_party/glm/simd/ |
D | common.h | 103 glm_vec4 const cmp1 = _mm_cmpgt_ps(x, zro0); in glm_vec4_sign() 161 glm_vec4 const cmp0 = _mm_cmpgt_ps(x, rnd0); in glm_vec4_ceil()
|
/external/eigen/unsupported/Eigen/src/MoreVectorization/ |
D | MathFunctions.h | 50 Packet4f branch_mask= _mm_cmpgt_ps(a, p4f_half);//this is to select which branch to take in pasin()
|
/external/libopus/celt/x86/ |
D | vq_sse2.c | 176 pos = _mm_max_epi16(pos, _mm_and_si128(count, _mm_castps_si128(_mm_cmpgt_ps(r4, max)))); in op_pvq_search_sse2()
|
/external/llvm-project/clang/test/CodeGen/ |
D | ppc-xmmintrin.c | 127 res = _mm_cmpgt_ps(m1, m2); in test_cmp()
|
/external/skia/include/private/ |
D | SkNx_sse.h | 33 __m128 too_big = _mm_cmpgt_ps(roundtrip, v); in emulate_mm_floor_ps() 86 AI SkNx operator > (const SkNx& o) const { return _mm_cmpgt_ps (fVec, o.fVec); } 177 AI SkNx operator > (const SkNx& o) const { return _mm_cmpgt_ps (fVec, o.fVec); }
|
/external/clang/test/CodeGen/ |
D | sse-builtins.c | 75 return _mm_cmpgt_ps(__a, __b); in test_mm_cmpgt_ps()
|
/external/skqp/include/private/ |
D | SkNx_sse.h | 33 __m128 too_big = _mm_cmpgt_ps(roundtrip, v); in emulate_mm_floor_ps() 86 AI SkNx operator > (const SkNx& o) const { return _mm_cmpgt_ps (fVec, o.fVec); } 179 AI SkNx operator > (const SkNx& o) const { return _mm_cmpgt_ps (fVec, o.fVec); }
|
/external/eigen/Eigen/src/Core/arch/SSE/ |
D | MathFunctions.h | 144 Packet4f mask = _mm_cmpgt_ps(tmp, fx);
|
/external/llvm-project/clang/lib/Headers/ |
D | xmmintrin.h | 632 _mm_cmpgt_ps(__m128 __a, __m128 __b) in _mm_cmpgt_ps() function
|
/external/clang/lib/Headers/ |
D | xmmintrin.h | 646 _mm_cmpgt_ps(__m128 __a, __m128 __b) in _mm_cmpgt_ps() function
|
/external/pffft/ |
D | sse2neon.h | 3837 FORCE_INLINE __m128 _mm_cmpgt_ps(__m128 a, __m128 b) in _mm_cmpgt_ps() function 3847 return _mm_move_ss(a, _mm_cmpgt_ps(a, b)); in _mm_cmpgt_ss() 3948 return _mm_cmpgt_ps(a, b); in _mm_cmpnle_ps()
|
/external/llvm-project/clang/lib/Headers/ppc_wrappers/ |
D | xmmintrin.h | 496 _mm_cmpgt_ps (__m128 __A, __m128 __B) in _mm_cmpgt_ps() function
|