| /external/pytorch/aten/src/ATen/native/quantized/cpu/qnnpack/src/requantization/ |
| D | fp32-scalar.c | 51 const int32_t y_clamped = (int32_t)( in pytorch_qnnp_requantize_fp32__scalar_lrintf() local 103 const float y_clamped = in pytorch_qnnp_requantize_fp32__scalar_magic() local
|
| D | precise-scalar.c | 134 const int32_t y_clamped = in pytorch_qnnp_requantize_precise__scalar_unsigned32() local 239 const int32_t y_clamped = in pytorch_qnnp_requantize_precise__scalar_unsigned64() local 339 const int32_t y_clamped = in pytorch_qnnp_requantize_precise__scalar_signed64() local
|
| D | fp32-psimd.c | 74 const psimd_f32 y_clamped = in pytorch_qnnp_requantize_fp32__psimd() local
|
| D | gemmlowp-scalar.c | 68 const int32_t y_clamped = in pytorch_qnnp_requantize_gemmlowp__scalar() local
|
| D | fp32-neon.c | 109 const float32x4_t y_clamped = vminq_f32(vmaxq_f32(y_scaled, vfmin), vfmax); in pytorch_qnnp_requantize_fp32__neon() local
|
| D | q31-scalar.c | 137 const int32_t y_clamped = in pytorch_qnnp_requantize_q31__scalar() local
|
| D | precise-psimd.c | 144 const psimd_u32 y_clamped = in pytorch_qnnp_requantize_precise__psimd() local
|
| /external/XNNPACK/src/qu8-requantization/ |
| D | fp32-scalar-fmagic.c | 48 const float y_clamped = math_min_f32(math_max_f32(y_scaled, fmin), fmax); in xnn_qu8_requantize_fp32__scalar_fmagic() local
|
| D | fp32-scalar-lrintf.c | 46 const float y_clamped = math_min_f32(math_max_f32(y_scaled, fmin), fmax); in xnn_qu8_requantize_fp32__scalar_lrintf() local
|
| D | fp32-wasmsimd.c | 59 const v128_t y_clamped = wasm_f32x4_min(wasm_f32x4_max(y_scaled, vfmin), vfmax); in xnn_qu8_requantize_fp32__wasmsimd() local
|
| D | rndna-scalar-signed64.c | 73 const int32_t y_clamped = math_min_s32(math_max_s32(y_scaled, smin), smax); in xnn_qu8_requantize_rndna__scalar_signed64() local
|
| D | rndna-scalar-unsigned64.c | 80 const int32_t y_clamped = math_min_s32(math_max_s32(y_scaled, smin), smax); in xnn_qu8_requantize_rndna__scalar_unsigned64() local
|
| D | fp32-neon.c | 91 const float32x4_t y_clamped = vminq_f32(vmaxq_f32(y_scaled, vfmin), vfmax); in xnn_qu8_requantize_fp32__neon() local
|
| D | gemmlowp-scalar.c | 108 const int32_t y_clamped = math_min_s32(math_max_s32(y_scaled, smin), smax); in xnn_qu8_requantize_gemmlowp__scalar() local
|
| D | rndna-scalar-unsigned32.c | 108 const int32_t y_clamped = math_min_s32(math_max_s32(y_scaled, smin), smax); in xnn_qu8_requantize_rndna__scalar_unsigned32() local
|
| /external/XNNPACK/src/qs8-requantization/ |
| D | fp32-scalar-fmagic.c | 48 const float y_clamped = math_min_f32(math_max_f32(y_scaled, fmin), fmax); in xnn_qs8_requantize_fp32__scalar_fmagic() local
|
| D | fp32-scalar-lrintf.c | 46 const float y_clamped = math_min_f32(math_max_f32(y_scaled, fmin), fmax); in xnn_qs8_requantize_fp32__scalar_lrintf() local
|
| D | fp32-wasmsimd.c | 59 const v128_t y_clamped = wasm_f32x4_min(wasm_f32x4_max(y_scaled, vfmin), vfmax); in xnn_qs8_requantize_fp32__wasmsimd() local
|
| D | rndnu-scalar.c | 67 const int32_t y_clamped = math_min_s32(math_max_s32(y_scaled, smin), smax); in xnn_qs8_requantize_rndnu__scalar() local
|
| D | rndna-scalar-signed64.c | 73 const int32_t y_clamped = math_min_s32(math_max_s32(y_scaled, smin), smax); in xnn_qs8_requantize_rndna__scalar_signed64() local
|
| D | rndna-scalar-unsigned64.c | 80 const int32_t y_clamped = math_min_s32(math_max_s32(y_scaled, smin), smax); in xnn_qs8_requantize_rndna__scalar_unsigned64() local
|
| D | fp32-neon.c | 115 const float32x4_t y_clamped = vminq_f32(vmaxq_f32(y_scaled, vfmin), vfmax); in xnn_qs8_requantize_fp32__neon() local
|
| D | gemmlowp-scalar.c | 108 const int32_t y_clamped = math_min_s32(math_max_s32(y_scaled, smin), smax); in xnn_qs8_requantize_gemmlowp__scalar() local
|
| D | rndna-scalar-unsigned32.c | 108 const int32_t y_clamped = math_min_s32(math_max_s32(y_scaled, smin), smax); in xnn_qs8_requantize_rndna__scalar_unsigned32() local
|