/external/XNNPACK/src/requantization/ |
D | fp32-scalar.c | 43 const float w_scaled = (float) w * scale; in xnn_requantize_fp32__scalar_lrintf() local 95 const float w_scaled = (float) w * scale; in xnn_requantize_fp32__scalar_magic() local
|
D | precise-scalar.c | 105 const int32_t w_scaled = (int32_t)(w >= 0 ? w_abs_scaled : -w_abs_scaled); in xnn_requantize_precise__scalar_unsigned32() local 190 const int32_t w_scaled = (int32_t)(w >= 0 ? w_abs_scaled : -w_abs_scaled); in xnn_requantize_precise__scalar_unsigned64() local 268 const int32_t w_scaled = (int32_t) asr_s64(w_adjusted_product + rounding, shift); in xnn_requantize_precise__scalar_signed64() local
|
D | precise-neon.c | 112 …const int32x4_t w_scaled = vuzp1q_s32(vreinterpretq_s32_s64(w01_scaled), vreinterpretq_s32_s64(w23… in xnn_requantize_precise__neon() local 121 const int32x4_t w_scaled = vcombine_s32(vmovn_s64(w01_scaled), vmovn_s64(w23_scaled)); in xnn_requantize_precise__neon() local
|
D | fp32-sse2.c | 50 const __m128 w_scaled = _mm_mul_ps(_mm_cvtepi32_ps(w), vscale); in xnn_requantize_fp32__sse2() local
|
D | gemmlowp-ssse3.c | 60 const __m128i w_scaled = gemmlowp_sse_rdivbypo2_s32(w_product, shift); in xnn_requantize_gemmlowp__ssse3() local
|
D | gemmlowp-sse4.c | 60 const __m128i w_scaled = gemmlowp_sse_rdivbypo2_s32(w_product, shift); in xnn_requantize_gemmlowp__sse4() local
|
D | gemmlowp-sse2.c | 60 const __m128i w_scaled = gemmlowp_sse_rdivbypo2_s32(w_product, shift); in xnn_requantize_gemmlowp__sse2() local
|
D | fp32-psimd.c | 52 const psimd_f32 w_scaled = psimd_cvt_s32_f32(w) * vscale; in xnn_requantize_fp32__psimd() local
|
D | gemmlowp-scalar.c | 57 const int32_t w_scaled = gemmlowp_scalar_rdivbypo2_s32(w_product, shift); in xnn_requantize_gemmlowp__scalar() local
|
D | q31-neon.c | 79 const int32x4_t w_scaled = vrshlq_s32(w_adjusted_product, vshift); in xnn_requantize_q31__neon() local
|
D | fp32-neon.c | 57 const float32x4_t w_scaled = vmulq_f32(vcvtq_f32_s32(w), vscale); in xnn_requantize_fp32__neon() local
|
D | q31-scalar.c | 105 const int32_t w_scaled = asr_s32(w_q31product, shift) + (int32_t)(w_remainder > threshold); in xnn_requantize_q31__scalar() local
|
D | gemmlowp-neon.c | 87 const int32x4_t w_scaled = vrshlq_s32(w_adjusted_product, vshift); in xnn_requantize_gemmlowp__neon() local
|
D | precise-sse4.c | 90 const __m128i w_scaled = _mm_sign_epi32(w_abs_scaled, w); in xnn_requantize_precise__sse4() local
|
D | q31-sse4.c | 105 const __m128i w_scaled = in xnn_requantize_q31__sse4() local
|
D | precise-ssse3.c | 98 const __m128i w_scaled = _mm_sign_epi32(w_abs_scaled, w); in xnn_requantize_precise__ssse3() local
|
D | precise-sse2.c | 103 const __m128i w_scaled = _mm_sub_epi32(_mm_xor_si128(w_abs_scaled, w_neg_mask), w_neg_mask); in xnn_requantize_precise__sse2() local
|
D | precise-psimd.c | 123 const psimd_s32 w_scaled = (psimd_s32)(w_abs_scaled ^ w_neg_mask) - w_neg_mask; in xnn_requantize_precise__psimd() local
|
D | q31-ssse3.c | 154 const __m128i w_scaled = in xnn_requantize_q31__ssse3() local
|
D | q31-sse2.c | 154 const __m128i w_scaled = in xnn_requantize_q31__sse2() local
|