Lines Matching refs:res_sat
2742 __m128i c7fffffff, res, res_sat, res_xor_a; in vqd_s32() local
2745 res_sat = _mm_srli_epi32(a, 31); in vqd_s32()
2746 res_sat = _mm_add_epi32(res_sat, c7fffffff); in vqd_s32()
2749 res_sat = _mm_and_si128(res_xor_a, res_sat); in vqd_s32()
2751 return _mm_or_si128(res, res_sat); in vqd_s32()
3296 __m128i c7fffffff, res, res_sat, res_xor_a, b_xor_a_; in vqaddq_s32() local
3299 res_sat = _mm_srli_epi32(a, 31); in vqaddq_s32()
3300 res_sat = _mm_add_epi32(res_sat, c7fffffff); in vqaddq_s32()
3305 res_sat = _mm_and_si128(res_xor_a, res_sat); in vqaddq_s32()
3307 return _mm_or_si128(res, res_sat); in vqaddq_s32()
4286 __m128i res_sat, mask, mask1; in _NEON2SSE_PERFORMANCE_WARNING() local
4289 res_sat = _mm_slli_epi64 (mul, 1); //double the result, saturation not considered in _NEON2SSE_PERFORMANCE_WARNING()
4290 mask1 = _mm_slli_epi64(res_sat, 32); //shift left then back right to in _NEON2SSE_PERFORMANCE_WARNING()
4292 mul = _mm_add_epi32 (res_sat, mask1); //actual rounding in _NEON2SSE_PERFORMANCE_WARNING()
4651 __m128i c7fffffff, res, res_sat, res_xor_a, b_xor_a; in vqsubq_s32() local
4654 res_sat = _mm_srli_epi32(a, 31); in vqsubq_s32()
4655 res_sat = _mm_add_epi32(res_sat, c7fffffff); in vqsubq_s32()
4660 res_sat = _mm_and_si128(res_xor_a, res_sat); in vqsubq_s32()
4662 return _mm_or_si128(res, res_sat); in vqsubq_s32()