Home
last modified time | relevance | path

Searched refs:_mm_cmpgt_epi32 (Results 1 – 25 of 45) sorted by relevance

12

/external/XNNPACK/src/requantization/
Dq31-ssse3.c66 const __m128i x_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), x); in xnn_requantize_q31__ssse3()
67 const __m128i y_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), y); in xnn_requantize_q31__ssse3()
68 const __m128i z_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), z); in xnn_requantize_q31__ssse3()
69 const __m128i w_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), w); in xnn_requantize_q31__ssse3()
140 …_mm_add_epi32(_mm_and_si128(x_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), x… in xnn_requantize_q31__ssse3()
142 …_mm_add_epi32(_mm_and_si128(y_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), y… in xnn_requantize_q31__ssse3()
144 …_mm_add_epi32(_mm_and_si128(z_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), z… in xnn_requantize_q31__ssse3()
146 …_mm_add_epi32(_mm_and_si128(w_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), w… in xnn_requantize_q31__ssse3()
149 … _mm_sub_epi32(_mm_sra_epi32(x_q31product, vshift), _mm_cmpgt_epi32(x_remainder, vthreshold)); in xnn_requantize_q31__ssse3()
151 … _mm_sub_epi32(_mm_sra_epi32(y_q31product, vshift), _mm_cmpgt_epi32(y_remainder, vthreshold)); in xnn_requantize_q31__ssse3()
[all …]
Dq31-sse2.c61 const __m128i x_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), x); in xnn_requantize_q31__sse2()
62 const __m128i y_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), y); in xnn_requantize_q31__sse2()
63 const __m128i z_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), z); in xnn_requantize_q31__sse2()
64 const __m128i w_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), w); in xnn_requantize_q31__sse2()
140 …_mm_add_epi32(_mm_and_si128(x_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), x… in xnn_requantize_q31__sse2()
142 …_mm_add_epi32(_mm_and_si128(y_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), y… in xnn_requantize_q31__sse2()
144 …_mm_add_epi32(_mm_and_si128(z_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), z… in xnn_requantize_q31__sse2()
146 …_mm_add_epi32(_mm_and_si128(w_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), w… in xnn_requantize_q31__sse2()
149 … _mm_sub_epi32(_mm_sra_epi32(x_q31product, vshift), _mm_cmpgt_epi32(x_remainder, vthreshold)); in xnn_requantize_q31__sse2()
151 … _mm_sub_epi32(_mm_sra_epi32(y_q31product, vshift), _mm_cmpgt_epi32(y_remainder, vthreshold)); in xnn_requantize_q31__sse2()
[all …]
Dq31-sse4.c91 …_mm_add_epi32(_mm_and_si128(x_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), x… in xnn_requantize_q31__sse4()
93 …_mm_add_epi32(_mm_and_si128(y_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), y… in xnn_requantize_q31__sse4()
95 …_mm_add_epi32(_mm_and_si128(z_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), z… in xnn_requantize_q31__sse4()
97 …_mm_add_epi32(_mm_and_si128(w_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), w… in xnn_requantize_q31__sse4()
100 … _mm_sub_epi32(_mm_sra_epi32(x_q31product, vshift), _mm_cmpgt_epi32(x_remainder, vthreshold)); in xnn_requantize_q31__sse4()
102 … _mm_sub_epi32(_mm_sra_epi32(y_q31product, vshift), _mm_cmpgt_epi32(y_remainder, vthreshold)); in xnn_requantize_q31__sse4()
104 … _mm_sub_epi32(_mm_sra_epi32(z_q31product, vshift), _mm_cmpgt_epi32(z_remainder, vthreshold)); in xnn_requantize_q31__sse4()
106 … _mm_sub_epi32(_mm_sra_epi32(w_q31product, vshift), _mm_cmpgt_epi32(w_remainder, vthreshold)); in xnn_requantize_q31__sse4()
Dprecise-sse2.c52 const __m128i x_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), x); in xnn_requantize_precise__sse2()
53 const __m128i y_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), y); in xnn_requantize_precise__sse2()
54 const __m128i z_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), z); in xnn_requantize_precise__sse2()
55 const __m128i w_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), w); in xnn_requantize_precise__sse2()
/external/XNNPACK/src/q8-vadd/
Dsse2.c60 …_mm_add_epi32(_mm_and_si128(vacc_lo, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_l… in xnn_q8_vadd_ukernel__sse2()
62 …_mm_add_epi32(_mm_and_si128(vacc_hi, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_h… in xnn_q8_vadd_ukernel__sse2()
64 …vacc_lo = _mm_sub_epi32(_mm_sra_epi32(vacc_lo, vshift), _mm_cmpgt_epi32(vrem_lo, vremainder_thresh… in xnn_q8_vadd_ukernel__sse2()
65 …vacc_hi = _mm_sub_epi32(_mm_sra_epi32(vacc_hi, vshift), _mm_cmpgt_epi32(vrem_hi, vremainder_thresh… in xnn_q8_vadd_ukernel__sse2()
102 …_mm_add_epi32(_mm_and_si128(vacc_lo, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_l… in xnn_q8_vadd_ukernel__sse2()
104 …_mm_add_epi32(_mm_and_si128(vacc_hi, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_h… in xnn_q8_vadd_ukernel__sse2()
106 …vacc_lo = _mm_sub_epi32(_mm_sra_epi32(vacc_lo, vshift), _mm_cmpgt_epi32(vrem_lo, vremainder_thresh… in xnn_q8_vadd_ukernel__sse2()
107 …vacc_hi = _mm_sub_epi32(_mm_sra_epi32(vacc_hi, vshift), _mm_cmpgt_epi32(vrem_hi, vremainder_thresh… in xnn_q8_vadd_ukernel__sse2()
/external/XNNPACK/src/f32-prelu/gen/
Dsse2-2x8.c62 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8()
64 …const __m128 vmask0x4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8()
66 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8()
68 …const __m128 vmask1x4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8()
102 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8()
104 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8()
130 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8()
132 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8()
Dsse2-2x4.c59 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x4()
61 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x4()
87 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x4()
89 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x4()
/external/libvpx/libvpx/vp9/encoder/x86/
Dvp9_highbd_block_error_intrin_sse2.c36 cmp0 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff, max), in vp9_highbd_block_error_sse2()
38 cmp1 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff2, max), in vp9_highbd_block_error_sse2()
40 cmp2 = _mm_xor_si128(_mm_cmpgt_epi32(mm_dqcoeff, max), in vp9_highbd_block_error_sse2()
42 cmp3 = _mm_xor_si128(_mm_cmpgt_epi32(mm_dqcoeff2, max), in vp9_highbd_block_error_sse2()
/external/libaom/libaom/av1/encoder/x86/
Dhighbd_block_error_intrin_sse2.c37 cmp0 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff, max), in av1_highbd_block_error_sse2()
39 cmp1 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff2, max), in av1_highbd_block_error_sse2()
41 cmp2 = _mm_xor_si128(_mm_cmpgt_epi32(mm_dqcoeff, max), in av1_highbd_block_error_sse2()
43 cmp3 = _mm_xor_si128(_mm_cmpgt_epi32(mm_dqcoeff2, max), in av1_highbd_block_error_sse2()
/external/XNNPACK/src/q8-gemm/
D4x4c2-sse2.c206 const __m128i vnmask0x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc0x0123); in xnn_q8_gemm_ukernel_4x4c2__sse2()
207 const __m128i vnmask1x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc1x0123); in xnn_q8_gemm_ukernel_4x4c2__sse2()
208 const __m128i vnmask2x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc2x0123); in xnn_q8_gemm_ukernel_4x4c2__sse2()
209 const __m128i vnmask3x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc3x0123); in xnn_q8_gemm_ukernel_4x4c2__sse2()
278 …_mm_add_epi32(_mm_and_si128(vq31prod0x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_4x4c2__sse2()
280 …_mm_add_epi32(_mm_and_si128(vq31prod1x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_4x4c2__sse2()
282 …_mm_add_epi32(_mm_and_si128(vq31prod2x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_4x4c2__sse2()
284 …_mm_add_epi32(_mm_and_si128(vq31prod3x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_4x4c2__sse2()
290 …_mm_sub_epi32(_mm_sra_epi32(vq31prod0x0123, vshift), _mm_cmpgt_epi32(vrem0x0123, vremainder_thresh… in xnn_q8_gemm_ukernel_4x4c2__sse2()
292 …_mm_sub_epi32(_mm_sra_epi32(vq31prod1x0123, vshift), _mm_cmpgt_epi32(vrem1x0123, vremainder_thresh… in xnn_q8_gemm_ukernel_4x4c2__sse2()
[all …]
D2x4c8-sse2.c110 const __m128i vnmask0x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc0x0123); in xnn_q8_gemm_ukernel_2x4c8__sse2()
111 const __m128i vnmask1x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc1x0123); in xnn_q8_gemm_ukernel_2x4c8__sse2()
154 …_mm_add_epi32(_mm_and_si128(vq31prod0x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_2x4c8__sse2()
156 …_mm_add_epi32(_mm_and_si128(vq31prod1x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_2x4c8__sse2()
162 …_mm_sub_epi32(_mm_sra_epi32(vq31prod0x0123, vshift), _mm_cmpgt_epi32(vrem0x0123, vremainder_thresh… in xnn_q8_gemm_ukernel_2x4c8__sse2()
164 …_mm_sub_epi32(_mm_sra_epi32(vq31prod1x0123, vshift), _mm_cmpgt_epi32(vrem1x0123, vremainder_thresh… in xnn_q8_gemm_ukernel_2x4c8__sse2()
/external/XNNPACK/src/q8-igemm/
D4x4c2-sse2.c186 const __m128i vnmask0x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc0x0123); in xnn_q8_igemm_ukernel_4x4c2__sse2()
187 const __m128i vnmask1x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc1x0123); in xnn_q8_igemm_ukernel_4x4c2__sse2()
188 const __m128i vnmask2x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc2x0123); in xnn_q8_igemm_ukernel_4x4c2__sse2()
189 const __m128i vnmask3x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc3x0123); in xnn_q8_igemm_ukernel_4x4c2__sse2()
258 …_mm_add_epi32(_mm_and_si128(vq31prod0x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_igemm_ukernel_4x4c2__sse2()
260 …_mm_add_epi32(_mm_and_si128(vq31prod1x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_igemm_ukernel_4x4c2__sse2()
262 …_mm_add_epi32(_mm_and_si128(vq31prod2x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_igemm_ukernel_4x4c2__sse2()
264 …_mm_add_epi32(_mm_and_si128(vq31prod3x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_igemm_ukernel_4x4c2__sse2()
269 …vacc0x0123 = _mm_sub_epi32(_mm_sra_epi32(vq31prod0x0123, vshift), _mm_cmpgt_epi32(vrem0x0123, vrem… in xnn_q8_igemm_ukernel_4x4c2__sse2()
270 …vacc1x0123 = _mm_sub_epi32(_mm_sra_epi32(vq31prod1x0123, vshift), _mm_cmpgt_epi32(vrem1x0123, vrem… in xnn_q8_igemm_ukernel_4x4c2__sse2()
[all …]
/external/XNNPACK/src/q8-dwconv/
Dup8x9-sse2.c132 const __m128i vnmask_lo0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_dwconv_ukernel_up8x9__sse2()
133 const __m128i vnmask_hi0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_dwconv_ukernel_up8x9__sse2()
176 …_mm_add_epi32(_mm_and_si128(vq31prod_lo0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128()… in xnn_q8_dwconv_ukernel_up8x9__sse2()
178 …_mm_add_epi32(_mm_and_si128(vq31prod_hi0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128()… in xnn_q8_dwconv_ukernel_up8x9__sse2()
183 …const __m128i vout_lo = _mm_sub_epi32(_mm_sra_epi32(vq31prod_lo0123, vshift), _mm_cmpgt_epi32(vrem… in xnn_q8_dwconv_ukernel_up8x9__sse2()
184 …const __m128i vout_hi = _mm_sub_epi32(_mm_sra_epi32(vq31prod_hi0123, vshift), _mm_cmpgt_epi32(vrem… in xnn_q8_dwconv_ukernel_up8x9__sse2()
282 const __m128i vnmask_lo0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_dwconv_ukernel_up8x9__sse2()
283 const __m128i vnmask_hi0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_dwconv_ukernel_up8x9__sse2()
326 …_mm_add_epi32(_mm_and_si128(vq31prod_lo0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128()… in xnn_q8_dwconv_ukernel_up8x9__sse2()
328 …_mm_add_epi32(_mm_and_si128(vq31prod_hi0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128()… in xnn_q8_dwconv_ukernel_up8x9__sse2()
[all …]
/external/XNNPACK/src/f32-sigmoid/gen/
Dsse2-p5-div-x24.c184 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
185 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
186 … __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89AB))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
187 … __m128 vmCDEF = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxCDEF))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
188 … __m128 vmGHIJ = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxGHIJ))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
189 … __m128 vmKLMN = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxKLMN))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
264 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
328 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
Dsse2-p5-div-x20.c168 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
169 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
170 … __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89AB))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
171 … __m128 vmCDEF = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxCDEF))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
172 … __m128 vmGHIJ = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxGHIJ))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
245 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
309 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
Dsse2-p5-div-x16.c152 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16()
153 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16()
154 … __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89AB))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16()
155 … __m128 vmCDEF = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxCDEF))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16()
226 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16()
290 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16()
Dsse2-p5-div-x12.c136 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12()
137 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12()
138 … __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89AB))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12()
207 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12()
271 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12()
Dsse2-p5-div-x8.c120 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8()
121 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8()
188 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8()
252 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8()
Dsse2-p5-div-x4.c99 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x4()
163 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x4()
/external/libaom/libaom/aom_dsp/x86/
Dhighbd_adaptive_quantize_sse2.c77 cmp_mask0 = _mm_cmpgt_epi32(coeff[0], threshold[0]); in highbd_update_mask0()
79 cmp_mask1 = _mm_cmpgt_epi32(coeff[1], threshold[1]); in highbd_update_mask0()
161 cmp_mask0 = _mm_cmpgt_epi32(qcoeff0, zbin); in aom_highbd_quantize_b_adaptive_sse2()
163 cmp_mask1 = _mm_cmpgt_epi32(qcoeff1, zbin); in aom_highbd_quantize_b_adaptive_sse2()
218 cmp_mask0 = _mm_cmpgt_epi32(qcoeff0, zbin); in aom_highbd_quantize_b_adaptive_sse2()
219 cmp_mask1 = _mm_cmpgt_epi32(qcoeff1, zbin); in aom_highbd_quantize_b_adaptive_sse2()
378 cmp_mask0 = _mm_cmpgt_epi32(qcoeff0, zbin); in aom_highbd_quantize_b_32x32_adaptive_sse2()
380 cmp_mask1 = _mm_cmpgt_epi32(qcoeff1, zbin); in aom_highbd_quantize_b_32x32_adaptive_sse2()
435 cmp_mask0 = _mm_cmpgt_epi32(qcoeff0, zbin); in aom_highbd_quantize_b_32x32_adaptive_sse2()
436 cmp_mask1 = _mm_cmpgt_epi32(qcoeff1, zbin); in aom_highbd_quantize_b_32x32_adaptive_sse2()
[all …]
Dquantize_x86.h176 cmp_mask0 = _mm_cmpgt_epi32(coeff[0], threshold[0]); in update_mask0()
178 cmp_mask1 = _mm_cmpgt_epi32(coeff[1], threshold[1]); in update_mask0()
180 cmp_mask2 = _mm_cmpgt_epi32(coeff[2], threshold[1]); in update_mask0()
182 cmp_mask3 = _mm_cmpgt_epi32(coeff[3], threshold[1]); in update_mask0()
Dhighbd_quantize_intrin_sse2.c50 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in aom_highbd_quantize_b_sse2()
69 tmp1 = _mm_cmpgt_epi32(coeffs, zbins[i != 0]); in aom_highbd_quantize_b_sse2()
123 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in aom_highbd_quantize_b_32x32_sse2()
181 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in aom_highbd_quantize_b_64x64_sse2()
/external/libvpx/libvpx/vpx_dsp/x86/
Dhighbd_quantize_intrin_sse2.c54 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in vpx_highbd_quantize_b_sse2()
73 tmp1 = _mm_cmpgt_epi32(coeffs, zbins[i != 0]); in vpx_highbd_quantize_b_sse2()
130 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in vpx_highbd_quantize_b_32x32_sse2()
/external/XNNPACK/src/q8-gavgpool/
Dup7-sse2.c89 const __m128i vneg_mask_lo = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_gavgpool_ukernel_up7__sse2()
90 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_gavgpool_ukernel_up7__sse2()
158 const __m128i vneg_mask_lo = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_gavgpool_ukernel_up7__sse2()
159 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_gavgpool_ukernel_up7__sse2()
/external/XNNPACK/src/q8-avgpool/
Dup9-sse2.c108 const __m128i vneg_mask_lo = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_avgpool_ukernel_up9__sse2()
109 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_avgpool_ukernel_up9__sse2()
183 const __m128i vneg_mask_lo = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_avgpool_ukernel_up9__sse2()
184 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_avgpool_ukernel_up9__sse2()

12