/external/tensorflow/tensorflow/lite/micro/kernels/ |
D | activations.cc | 57 int32_t clamped = in ReluQuantized() local 62 clamped = std::max(data.params.quantized_activation_min, clamped); in ReluQuantized() 63 clamped = std::min(data.params.quantized_activation_max, clamped); in ReluQuantized() 64 output_data[i] = static_cast<T>(clamped); in ReluQuantized() 103 const float clamped = val < lower ? lower : val; in ReluFloat() local 104 output_data[i] = clamped; in ReluFloat() 115 const float clamped = val > upper ? upper : val < lower ? lower : val; in Relu6Float() local 116 output_data[i] = clamped; in Relu6Float() 127 const Q clamped = val > upper ? upper : val < lower ? lower : val; in Relu6Quantized() local 128 output_data[i] = clamped; in Relu6Quantized()
|
/external/webrtc/api/audio/ |
D | echo_canceller3_config.cc | 21 float clamped = rtc::SafeClamp(*value, min, max); in Limit() local 22 clamped = std::isfinite(clamped) ? clamped : min; in Limit() 23 bool res = *value == clamped; in Limit() 24 *value = clamped; in Limit() 29 size_t clamped = rtc::SafeClamp(*value, min, max); in Limit() local 30 bool res = *value == clamped; in Limit() 31 *value = clamped; in Limit() 36 int clamped = rtc::SafeClamp(*value, min, max); in Limit() local 37 bool res = *value == clamped; in Limit() 38 *value = clamped; in Limit() [all …]
|
/external/angle/src/compiler/translator/tree_ops/ |
D | ClampIndirectIndices.cpp | 102 TIntermTyped *clamped = in visitBinary() local 108 TIntermSequence constructorArgs = {clamped}; in visitBinary() 109 clamped = TIntermAggregate::CreateConstructor(*StaticType::GetBasic<EbtInt>(), in visitBinary() 114 queueReplacementWithParent(node, node->getRight(), clamped, OriginalNode::IS_DROPPED); in visitBinary()
|
/external/mesa3d/src/compiler/nir/ |
D | nir_format_convert.h | 364 nir_ssa_def *clamped = nir_fmax(b, color, nir_imm_float(b, 0)); in nir_format_pack_11f11f10f() local 367 nir_ssa_def *p1 = nir_pack_half_2x16_split(b, nir_channel(b, clamped, 0), in nir_format_pack_11f11f10f() 368 nir_channel(b, clamped, 1)); in nir_format_pack_11f11f10f() 369 nir_ssa_def *p2 = nir_pack_half_2x16_split(b, nir_channel(b, clamped, 2), in nir_format_pack_11f11f10f() 390 nir_ssa_def *clamped = nir_fmin(b, color, nir_imm_float(b, MAX_RGB9E5)); in nir_format_pack_r9g9b9e5() local 393 clamped = nir_bcsel(b, nir_ult(b, nir_imm_int(b, 0x7f800000), color), in nir_format_pack_r9g9b9e5() 394 nir_imm_float(b, 0), clamped); in nir_format_pack_r9g9b9e5() 397 nir_ssa_def *maxu = nir_umax(b, nir_channel(b, clamped, 0), in nir_format_pack_r9g9b9e5() 398 nir_umax(b, nir_channel(b, clamped, 1), in nir_format_pack_r9g9b9e5() 399 nir_channel(b, clamped, 2))); in nir_format_pack_r9g9b9e5() [all …]
|
/external/webrtc/third_party/abseil-cpp/absl/base/internal/ |
D | spinlock.cc | 208 uint32_t clamped = static_cast<uint32_t>( in EncodeWaitCycles() local 211 if (clamped == 0) { in EncodeWaitCycles() 217 if (clamped == kSpinLockSleeper) { in EncodeWaitCycles() 220 return clamped; in EncodeWaitCycles()
|
/external/rust/crates/grpcio-sys/grpc/third_party/abseil-cpp/absl/base/internal/ |
D | spinlock.cc | 196 uint32_t clamped = static_cast<uint32_t>( in EncodeWaitCycles() local 199 if (clamped == 0) { in EncodeWaitCycles() 205 if (clamped == kSpinLockSleeper) { in EncodeWaitCycles() 208 return clamped; in EncodeWaitCycles()
|
/external/openscreen/third_party/abseil/src/absl/base/internal/ |
D | spinlock.cc | 196 uint32_t clamped = static_cast<uint32_t>( in EncodeWaitCycles() local 199 if (clamped == 0) { in EncodeWaitCycles() 205 if (clamped == kSpinLockSleeper) { in EncodeWaitCycles() 208 return clamped; in EncodeWaitCycles()
|
/external/abseil-cpp/absl/base/internal/ |
D | spinlock.cc | 208 uint32_t clamped = static_cast<uint32_t>( in EncodeWaitCycles() local 211 if (clamped == 0) { in EncodeWaitCycles() 217 if (clamped == kSpinLockSleeper) { in EncodeWaitCycles() 220 return clamped; in EncodeWaitCycles()
|
/external/angle/third_party/abseil-cpp/absl/base/internal/ |
D | spinlock.cc | 205 uint32_t clamped = static_cast<uint32_t>( in EncodeWaitCycles() local 208 if (clamped == 0) { in EncodeWaitCycles() 214 if (clamped == kSpinLockSleeper) { in EncodeWaitCycles() 217 return clamped; in EncodeWaitCycles()
|
/external/libtextclassifier/abseil-cpp/absl/base/internal/ |
D | spinlock.cc | 196 uint32_t clamped = static_cast<uint32_t>( in EncodeWaitCycles() local 199 if (clamped == 0) { in EncodeWaitCycles() 205 if (clamped == kSpinLockSleeper) { in EncodeWaitCycles() 208 return clamped; in EncodeWaitCycles()
|
/external/grpc-grpc/third_party/nanopb/ |
D | pb_decode.c | 1155 int64_t clamped; in pb_dec_varint() local 1172 clamped = *(int64_t*)dest = svalue; in pb_dec_varint() 1174 clamped = *(int32_t*)dest = (int32_t)svalue; in pb_dec_varint() 1176 clamped = *(int_least16_t*)dest = (int_least16_t)svalue; in pb_dec_varint() 1178 clamped = *(int_least8_t*)dest = (int_least8_t)svalue; in pb_dec_varint() 1182 if (clamped != svalue) in pb_dec_varint() 1190 uint64_t value, clamped; in pb_dec_uvarint() local 1196 clamped = *(uint64_t*)dest = value; in pb_dec_uvarint() 1198 clamped = *(uint32_t*)dest = (uint32_t)value; in pb_dec_uvarint() 1200 clamped = *(uint_least16_t*)dest = (uint_least16_t)value; in pb_dec_uvarint() [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | llvm.AMDGPU.rsq.clamped.ll | 5 ; FIXME: Uses of this should be moved to llvm.amdgcn.rsq.clamped, and 8 declare float @llvm.AMDGPU.rsq.clamped.f32(float) nounwind readnone 22 %rsq_clamped = call float @llvm.AMDGPU.rsq.clamped.f32(float %src) nounwind readnone
|
D | llvm.AMDGPU.rsq.clamped.f64.ll | 4 declare double @llvm.AMDGPU.rsq.clamped.f64(double) nounwind readnone 18 %rsq_clamped = call double @llvm.AMDGPU.rsq.clamped.f64(double %src) nounwind readnone
|
/external/nanopb-c/ |
D | pb_decode.c | 1293 pb_int64_t clamped; in pb_dec_varint() local 1310 clamped = *(pb_int64_t*)dest = svalue; in pb_dec_varint() 1312 clamped = *(int32_t*)dest = (int32_t)svalue; in pb_dec_varint() 1314 clamped = *(int_least16_t*)dest = (int_least16_t)svalue; in pb_dec_varint() 1316 clamped = *(int_least8_t*)dest = (int_least8_t)svalue; in pb_dec_varint() 1320 if (clamped != svalue) in pb_dec_varint() 1328 pb_uint64_t value, clamped; in pb_dec_uvarint() local 1334 clamped = *(pb_uint64_t*)dest = value; in pb_dec_uvarint() 1336 clamped = *(uint32_t*)dest = (uint32_t)value; in pb_dec_uvarint() 1338 clamped = *(uint_least16_t*)dest = (uint_least16_t)value; in pb_dec_uvarint() [all …]
|
/external/tensorflow/tensorflow/core/kernels/ |
D | fake_quant_ops_functor.h | 92 auto clamped = inputs.cwiseMin(nudged_max).cwiseMax(nudged_min); in operator() local 93 auto clamped_shifted = clamped - nudged_min; in operator() 143 const auto clamped = inputs.cwiseMin(nudged_max).cwiseMax(nudged_min); in operator() local 144 const auto clamped_shifted = clamped - nudged_min; in operator() 215 const auto clamped = in operator() local 217 const auto clamped_shifted = clamped - nudged_min; in operator()
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/ |
D | llvm.r600.recipsqrt.clamped.ll | 3 declare float @llvm.r600.recipsqrt.clamped.f32(float) nounwind readnone 8 %rsq_clamped = call float @llvm.r600.recipsqrt.clamped.f32(float %src)
|
/external/OpenCL-CTS/test_conformance/images/kernel_read_write/ |
D | test_common.h | 63 bool clampingErr = false, clamped = false, otherClampingBug = false; in determine_validation_error_offset() local 69 clamped = get_integer_coords_offset(x, y, z, xAddressOffset, yAddressOffset, in determine_validation_error_offset() 77 if (clamped && clampedX == actualX && clampedY == actualY in determine_validation_error_offset() 111 if (clamped && !otherClampingBug) in determine_validation_error_offset()
|
/external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
D | quantize.h | 47 int32_t clamped = std::min(std::max(unclamped, min_val), max_val); in AffineQuantize() local 48 output_data[i] = clamped; in AffineQuantize()
|
/external/mesa3d/src/panfrost/midgard/ |
D | midgard_errata_lod.c | 74 nir_ssa_def *clamped = nir_fmin(b, in nir_lod_errata_instr() local 78 nir_src_for_ssa(clamped)); in nir_lod_errata_instr()
|
/external/llvm-project/mlir/include/mlir/Dialect/Quant/ |
D | UniformSupport.h | 138 const double clamped = in quantizeF32ToInt8() local 143 int64_t clampedInt = static_cast<int8_t>(clamped); in quantizeF32ToInt8() 146 signlessResult = static_cast<uint8_t>(clamped); in quantizeF32ToInt8()
|
/external/mesa3d/docs/_extra/specs/ |
D | MESA_texture_signed_rgba.spec | 118 a framebuffer clamped to [0,1] all converted numbers will be in the range 146 are clamped to [-1,1] unless stated otherwise as in the case 147 of explicitly clamped to [0,1] for GL_COMBINE. or clamping the 158 requires texture source colors to be clamped to [0,1]. Texture 160 negative values generated post-filtering would be clamped to
|
/external/tensorflow/tensorflow/lite/kernels/internal/optimized/integer_ops/ |
D | mul.h | 119 const auto clamped = vmaxq_s8(output_activation_min_vector, in MulElementwise() local 121 vst1q_s8(output_data + i, clamped); in MulElementwise() 212 const auto clamped = vmaxq_s8(output_activation_min_vector, in MulSimpleBroadcast() local 214 vst1q_s8(output_data + i, clamped); in MulSimpleBroadcast()
|
D | add.h | 136 const int8x16_t clamped = in AddElementwise() local 139 vst1q_s8(output_data + i, clamped); in AddElementwise() 232 const int8x8_t clamped = in AddScalarBroadcast() local 235 vst1_s8(output_data + i, clamped); in AddScalarBroadcast()
|
/external/mesa3d/src/broadcom/compiler/ |
D | v3d_nir_lower_line_smooth.c | 130 nir_ssa_def *clamped = nir_fmin(&b, coverage, nir_imm_float(&b, 1.0f)); in initialise_coverage_var() local 132 nir_store_var(&b, state->coverage, clamped, 0x1 /* writemask */); in initialise_coverage_var()
|
/external/python/cpython2/Lib/test/decimaltestdata/ |
D | clamp.decTest | 2 -- clamp.decTest -- clamped exponent tests (format-independent) -- 129 -- clamped zeros... 144 -- clamped zeros... 187 -- subnormals clamped to 0-Etiny
|