/external/XNNPACK/src/requantization/ |
D | q31-ssse3.c | 66 const __m128i x_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), x); in xnn_requantize_q31__ssse3() 67 const __m128i y_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), y); in xnn_requantize_q31__ssse3() 68 const __m128i z_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), z); in xnn_requantize_q31__ssse3() 69 const __m128i w_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), w); in xnn_requantize_q31__ssse3() 140 …_mm_add_epi32(_mm_and_si128(x_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), x… in xnn_requantize_q31__ssse3() 142 …_mm_add_epi32(_mm_and_si128(y_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), y… in xnn_requantize_q31__ssse3() 144 …_mm_add_epi32(_mm_and_si128(z_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), z… in xnn_requantize_q31__ssse3() 146 …_mm_add_epi32(_mm_and_si128(w_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), w… in xnn_requantize_q31__ssse3() 149 … _mm_sub_epi32(_mm_sra_epi32(x_q31product, vshift), _mm_cmpgt_epi32(x_remainder, vthreshold)); in xnn_requantize_q31__ssse3() 151 … _mm_sub_epi32(_mm_sra_epi32(y_q31product, vshift), _mm_cmpgt_epi32(y_remainder, vthreshold)); in xnn_requantize_q31__ssse3() [all …]
|
D | q31-sse2.c | 61 const __m128i x_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), x); in xnn_requantize_q31__sse2() 62 const __m128i y_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), y); in xnn_requantize_q31__sse2() 63 const __m128i z_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), z); in xnn_requantize_q31__sse2() 64 const __m128i w_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), w); in xnn_requantize_q31__sse2() 140 …_mm_add_epi32(_mm_and_si128(x_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), x… in xnn_requantize_q31__sse2() 142 …_mm_add_epi32(_mm_and_si128(y_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), y… in xnn_requantize_q31__sse2() 144 …_mm_add_epi32(_mm_and_si128(z_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), z… in xnn_requantize_q31__sse2() 146 …_mm_add_epi32(_mm_and_si128(w_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), w… in xnn_requantize_q31__sse2() 149 … _mm_sub_epi32(_mm_sra_epi32(x_q31product, vshift), _mm_cmpgt_epi32(x_remainder, vthreshold)); in xnn_requantize_q31__sse2() 151 … _mm_sub_epi32(_mm_sra_epi32(y_q31product, vshift), _mm_cmpgt_epi32(y_remainder, vthreshold)); in xnn_requantize_q31__sse2() [all …]
|
D | q31-sse4.c | 91 …_mm_add_epi32(_mm_and_si128(x_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), x… in xnn_requantize_q31__sse4() 93 …_mm_add_epi32(_mm_and_si128(y_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), y… in xnn_requantize_q31__sse4() 95 …_mm_add_epi32(_mm_and_si128(z_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), z… in xnn_requantize_q31__sse4() 97 …_mm_add_epi32(_mm_and_si128(w_q31product, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), w… in xnn_requantize_q31__sse4() 100 … _mm_sub_epi32(_mm_sra_epi32(x_q31product, vshift), _mm_cmpgt_epi32(x_remainder, vthreshold)); in xnn_requantize_q31__sse4() 102 … _mm_sub_epi32(_mm_sra_epi32(y_q31product, vshift), _mm_cmpgt_epi32(y_remainder, vthreshold)); in xnn_requantize_q31__sse4() 104 … _mm_sub_epi32(_mm_sra_epi32(z_q31product, vshift), _mm_cmpgt_epi32(z_remainder, vthreshold)); in xnn_requantize_q31__sse4() 106 … _mm_sub_epi32(_mm_sra_epi32(w_q31product, vshift), _mm_cmpgt_epi32(w_remainder, vthreshold)); in xnn_requantize_q31__sse4()
|
D | precise-sse2.c | 52 const __m128i x_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), x); in xnn_requantize_precise__sse2() 53 const __m128i y_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), y); in xnn_requantize_precise__sse2() 54 const __m128i z_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), z); in xnn_requantize_precise__sse2() 55 const __m128i w_neg_mask = _mm_cmpgt_epi32(_mm_setzero_si128(), w); in xnn_requantize_precise__sse2()
|
/external/XNNPACK/src/q8-vadd/ |
D | sse2.c | 60 …_mm_add_epi32(_mm_and_si128(vacc_lo, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_l… in xnn_q8_vadd_ukernel__sse2() 62 …_mm_add_epi32(_mm_and_si128(vacc_hi, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_h… in xnn_q8_vadd_ukernel__sse2() 64 …vacc_lo = _mm_sub_epi32(_mm_sra_epi32(vacc_lo, vshift), _mm_cmpgt_epi32(vrem_lo, vremainder_thresh… in xnn_q8_vadd_ukernel__sse2() 65 …vacc_hi = _mm_sub_epi32(_mm_sra_epi32(vacc_hi, vshift), _mm_cmpgt_epi32(vrem_hi, vremainder_thresh… in xnn_q8_vadd_ukernel__sse2() 102 …_mm_add_epi32(_mm_and_si128(vacc_lo, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_l… in xnn_q8_vadd_ukernel__sse2() 104 …_mm_add_epi32(_mm_and_si128(vacc_hi, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_h… in xnn_q8_vadd_ukernel__sse2() 106 …vacc_lo = _mm_sub_epi32(_mm_sra_epi32(vacc_lo, vshift), _mm_cmpgt_epi32(vrem_lo, vremainder_thresh… in xnn_q8_vadd_ukernel__sse2() 107 …vacc_hi = _mm_sub_epi32(_mm_sra_epi32(vacc_hi, vshift), _mm_cmpgt_epi32(vrem_hi, vremainder_thresh… in xnn_q8_vadd_ukernel__sse2()
|
/external/XNNPACK/src/f32-prelu/gen/ |
D | sse2-2x8.c | 62 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8() 64 …const __m128 vmask0x4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8() 66 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8() 68 …const __m128 vmask1x4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8() 102 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8() 104 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8() 130 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8() 132 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x8()
|
D | sse2-2x4.c | 59 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x4() 61 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x4() 87 …const __m128 vmask0x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x4() 89 …const __m128 vmask1x0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(… in xnn_f32_prelu_ukernel__sse2_2x4()
|
/external/libvpx/libvpx/vp9/encoder/x86/ |
D | vp9_highbd_block_error_intrin_sse2.c | 36 cmp0 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff, max), in vp9_highbd_block_error_sse2() 38 cmp1 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff2, max), in vp9_highbd_block_error_sse2() 40 cmp2 = _mm_xor_si128(_mm_cmpgt_epi32(mm_dqcoeff, max), in vp9_highbd_block_error_sse2() 42 cmp3 = _mm_xor_si128(_mm_cmpgt_epi32(mm_dqcoeff2, max), in vp9_highbd_block_error_sse2()
|
/external/libaom/libaom/av1/encoder/x86/ |
D | highbd_block_error_intrin_sse2.c | 37 cmp0 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff, max), in av1_highbd_block_error_sse2() 39 cmp1 = _mm_xor_si128(_mm_cmpgt_epi32(mm_coeff2, max), in av1_highbd_block_error_sse2() 41 cmp2 = _mm_xor_si128(_mm_cmpgt_epi32(mm_dqcoeff, max), in av1_highbd_block_error_sse2() 43 cmp3 = _mm_xor_si128(_mm_cmpgt_epi32(mm_dqcoeff2, max), in av1_highbd_block_error_sse2()
|
/external/XNNPACK/src/q8-gemm/ |
D | 4x4c2-sse2.c | 206 const __m128i vnmask0x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc0x0123); in xnn_q8_gemm_ukernel_4x4c2__sse2() 207 const __m128i vnmask1x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc1x0123); in xnn_q8_gemm_ukernel_4x4c2__sse2() 208 const __m128i vnmask2x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc2x0123); in xnn_q8_gemm_ukernel_4x4c2__sse2() 209 const __m128i vnmask3x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc3x0123); in xnn_q8_gemm_ukernel_4x4c2__sse2() 278 …_mm_add_epi32(_mm_and_si128(vq31prod0x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_4x4c2__sse2() 280 …_mm_add_epi32(_mm_and_si128(vq31prod1x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_4x4c2__sse2() 282 …_mm_add_epi32(_mm_and_si128(vq31prod2x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_4x4c2__sse2() 284 …_mm_add_epi32(_mm_and_si128(vq31prod3x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_4x4c2__sse2() 290 …_mm_sub_epi32(_mm_sra_epi32(vq31prod0x0123, vshift), _mm_cmpgt_epi32(vrem0x0123, vremainder_thresh… in xnn_q8_gemm_ukernel_4x4c2__sse2() 292 …_mm_sub_epi32(_mm_sra_epi32(vq31prod1x0123, vshift), _mm_cmpgt_epi32(vrem1x0123, vremainder_thresh… in xnn_q8_gemm_ukernel_4x4c2__sse2() [all …]
|
D | 2x4c8-sse2.c | 110 const __m128i vnmask0x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc0x0123); in xnn_q8_gemm_ukernel_2x4c8__sse2() 111 const __m128i vnmask1x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc1x0123); in xnn_q8_gemm_ukernel_2x4c8__sse2() 154 …_mm_add_epi32(_mm_and_si128(vq31prod0x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_2x4c8__sse2() 156 …_mm_add_epi32(_mm_and_si128(vq31prod1x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_gemm_ukernel_2x4c8__sse2() 162 …_mm_sub_epi32(_mm_sra_epi32(vq31prod0x0123, vshift), _mm_cmpgt_epi32(vrem0x0123, vremainder_thresh… in xnn_q8_gemm_ukernel_2x4c8__sse2() 164 …_mm_sub_epi32(_mm_sra_epi32(vq31prod1x0123, vshift), _mm_cmpgt_epi32(vrem1x0123, vremainder_thresh… in xnn_q8_gemm_ukernel_2x4c8__sse2()
|
/external/XNNPACK/src/q8-igemm/ |
D | 4x4c2-sse2.c | 186 const __m128i vnmask0x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc0x0123); in xnn_q8_igemm_ukernel_4x4c2__sse2() 187 const __m128i vnmask1x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc1x0123); in xnn_q8_igemm_ukernel_4x4c2__sse2() 188 const __m128i vnmask2x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc2x0123); in xnn_q8_igemm_ukernel_4x4c2__sse2() 189 const __m128i vnmask3x0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc3x0123); in xnn_q8_igemm_ukernel_4x4c2__sse2() 258 …_mm_add_epi32(_mm_and_si128(vq31prod0x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_igemm_ukernel_4x4c2__sse2() 260 …_mm_add_epi32(_mm_and_si128(vq31prod1x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_igemm_ukernel_4x4c2__sse2() 262 …_mm_add_epi32(_mm_and_si128(vq31prod2x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_igemm_ukernel_4x4c2__sse2() 264 …_mm_add_epi32(_mm_and_si128(vq31prod3x0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128(),… in xnn_q8_igemm_ukernel_4x4c2__sse2() 269 …vacc0x0123 = _mm_sub_epi32(_mm_sra_epi32(vq31prod0x0123, vshift), _mm_cmpgt_epi32(vrem0x0123, vrem… in xnn_q8_igemm_ukernel_4x4c2__sse2() 270 …vacc1x0123 = _mm_sub_epi32(_mm_sra_epi32(vq31prod1x0123, vshift), _mm_cmpgt_epi32(vrem1x0123, vrem… in xnn_q8_igemm_ukernel_4x4c2__sse2() [all …]
|
/external/XNNPACK/src/q8-dwconv/ |
D | up8x9-sse2.c | 132 const __m128i vnmask_lo0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_dwconv_ukernel_up8x9__sse2() 133 const __m128i vnmask_hi0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_dwconv_ukernel_up8x9__sse2() 176 …_mm_add_epi32(_mm_and_si128(vq31prod_lo0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128()… in xnn_q8_dwconv_ukernel_up8x9__sse2() 178 …_mm_add_epi32(_mm_and_si128(vq31prod_hi0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128()… in xnn_q8_dwconv_ukernel_up8x9__sse2() 183 …const __m128i vout_lo = _mm_sub_epi32(_mm_sra_epi32(vq31prod_lo0123, vshift), _mm_cmpgt_epi32(vrem… in xnn_q8_dwconv_ukernel_up8x9__sse2() 184 …const __m128i vout_hi = _mm_sub_epi32(_mm_sra_epi32(vq31prod_hi0123, vshift), _mm_cmpgt_epi32(vrem… in xnn_q8_dwconv_ukernel_up8x9__sse2() 282 const __m128i vnmask_lo0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_dwconv_ukernel_up8x9__sse2() 283 const __m128i vnmask_hi0123 = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_dwconv_ukernel_up8x9__sse2() 326 …_mm_add_epi32(_mm_and_si128(vq31prod_lo0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128()… in xnn_q8_dwconv_ukernel_up8x9__sse2() 328 …_mm_add_epi32(_mm_and_si128(vq31prod_hi0123, vremainder_mask), _mm_cmpgt_epi32(_mm_setzero_si128()… in xnn_q8_dwconv_ukernel_up8x9__sse2() [all …]
|
/external/XNNPACK/src/f32-sigmoid/gen/ |
D | sse2-p5-div-x24.c | 184 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 185 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 186 … __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89AB))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 187 … __m128 vmCDEF = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxCDEF))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 188 … __m128 vmGHIJ = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxGHIJ))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 189 … __m128 vmKLMN = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxKLMN))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 264 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24() 328 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x24()
|
D | sse2-p5-div-x20.c | 168 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 169 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 170 … __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89AB))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 171 … __m128 vmCDEF = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxCDEF))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 172 … __m128 vmGHIJ = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxGHIJ))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 245 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20() 309 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x20()
|
D | sse2-p5-div-x16.c | 152 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 153 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 154 … __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89AB))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 155 … __m128 vmCDEF = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vxCDEF))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 226 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16() 290 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x16()
|
D | sse2-p5-div-x12.c | 136 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12() 137 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12() 138 … __m128 vm89AB = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx89AB))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12() 207 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12() 271 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x12()
|
D | sse2-p5-div-x8.c | 120 … __m128 vm0123 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx0123))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8() 121 … __m128 vm4567 = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx4567))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8() 188 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8() 252 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x8()
|
D | sse2-p5-div-x4.c | 99 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x4() 163 __m128 vm = _mm_castsi128_ps(_mm_cmpgt_epi32(_mm_setzero_si128(), _mm_castps_si128(vx))); in xnn_f32_sigmoid_ukernel__sse2_p5_div_x4()
|
/external/libaom/libaom/aom_dsp/x86/ |
D | highbd_adaptive_quantize_sse2.c | 77 cmp_mask0 = _mm_cmpgt_epi32(coeff[0], threshold[0]); in highbd_update_mask0() 79 cmp_mask1 = _mm_cmpgt_epi32(coeff[1], threshold[1]); in highbd_update_mask0() 161 cmp_mask0 = _mm_cmpgt_epi32(qcoeff0, zbin); in aom_highbd_quantize_b_adaptive_sse2() 163 cmp_mask1 = _mm_cmpgt_epi32(qcoeff1, zbin); in aom_highbd_quantize_b_adaptive_sse2() 218 cmp_mask0 = _mm_cmpgt_epi32(qcoeff0, zbin); in aom_highbd_quantize_b_adaptive_sse2() 219 cmp_mask1 = _mm_cmpgt_epi32(qcoeff1, zbin); in aom_highbd_quantize_b_adaptive_sse2() 378 cmp_mask0 = _mm_cmpgt_epi32(qcoeff0, zbin); in aom_highbd_quantize_b_32x32_adaptive_sse2() 380 cmp_mask1 = _mm_cmpgt_epi32(qcoeff1, zbin); in aom_highbd_quantize_b_32x32_adaptive_sse2() 435 cmp_mask0 = _mm_cmpgt_epi32(qcoeff0, zbin); in aom_highbd_quantize_b_32x32_adaptive_sse2() 436 cmp_mask1 = _mm_cmpgt_epi32(qcoeff1, zbin); in aom_highbd_quantize_b_32x32_adaptive_sse2() [all …]
|
D | quantize_x86.h | 176 cmp_mask0 = _mm_cmpgt_epi32(coeff[0], threshold[0]); in update_mask0() 178 cmp_mask1 = _mm_cmpgt_epi32(coeff[1], threshold[1]); in update_mask0() 180 cmp_mask2 = _mm_cmpgt_epi32(coeff[2], threshold[1]); in update_mask0() 182 cmp_mask3 = _mm_cmpgt_epi32(coeff[3], threshold[1]); in update_mask0()
|
D | highbd_quantize_intrin_sse2.c | 50 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in aom_highbd_quantize_b_sse2() 69 tmp1 = _mm_cmpgt_epi32(coeffs, zbins[i != 0]); in aom_highbd_quantize_b_sse2() 123 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in aom_highbd_quantize_b_32x32_sse2() 181 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in aom_highbd_quantize_b_64x64_sse2()
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | highbd_quantize_intrin_sse2.c | 54 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in vpx_highbd_quantize_b_sse2() 73 tmp1 = _mm_cmpgt_epi32(coeffs, zbins[i != 0]); in vpx_highbd_quantize_b_sse2() 130 cmp2 = _mm_cmpgt_epi32(coeffs, nzbins[i != 0]); in vpx_highbd_quantize_b_32x32_sse2()
|
/external/XNNPACK/src/q8-gavgpool/ |
D | up7-sse2.c | 89 const __m128i vneg_mask_lo = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_gavgpool_ukernel_up7__sse2() 90 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_gavgpool_ukernel_up7__sse2() 158 const __m128i vneg_mask_lo = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_gavgpool_ukernel_up7__sse2() 159 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_gavgpool_ukernel_up7__sse2()
|
/external/XNNPACK/src/q8-avgpool/ |
D | up9-sse2.c | 108 const __m128i vneg_mask_lo = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_avgpool_ukernel_up9__sse2() 109 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_avgpool_ukernel_up9__sse2() 183 const __m128i vneg_mask_lo = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_lo); in xnn_q8_avgpool_ukernel_up9__sse2() 184 const __m128i vneg_mask_hi = _mm_cmpgt_epi32(_mm_setzero_si128(), vacc_hi); in xnn_q8_avgpool_ukernel_up9__sse2()
|