/external/tensorflow/tensorflow/lite/kernels/internal/ |
D | per_channel_dequantize_test.cc | 30 const std::vector<float> scales = {0.5, 0.25}; in TEST() local 52 const std::vector<float> scales = {0.5, 0.25, 0.5, 0.25, 1.0}; in TEST() local 74 const std::vector<float> scales = {0.5, 0.25}; in TEST() local 98 const std::vector<float> scales = {0.5, 0.25, 0.5, 0.25, 1.0}; in TEST() local
|
D | per_channel_quantize_test.cc | 28 const std::vector<float> scales = {0.5, 0.25}; in TEST() local 49 const std::vector<float> scales = {0.5, 0.25, 0.5, 0.25, 1.0}; in TEST() local 70 const std::vector<float> scales = {0.5, 0.25}; in TEST() local 94 const std::vector<float> scales = {0.5, 0.25, 0.5, 0.25, 1.0}; in TEST() local
|
/external/libtextclassifier/native/annotator/ |
D | quantization_test.cc | 40 std::vector<float> scales{{0.1, 9.0, -7.0}}; in TEST() local 92 std::vector<float> scales(num_buckets); in TEST() local 115 std::vector<float> scales(num_buckets, 1.0); in TEST() local 135 std::vector<float> scales(num_buckets, 1.0); in TEST() local
|
D | quantization.cc | 28 void DequantizeAdd8bit(const float* scales, const uint8* embeddings, in DequantizeAdd8bit() 40 void DequantizeAddNBit(const float* scales, const uint8* embeddings, in DequantizeAddNBit() 73 bool DequantizeAdd(const float* scales, const uint8* embeddings, in DequantizeAdd()
|
D | model-executor.cc | 71 const TfLiteTensor* scales = interpreter->tensor(1); in FromBuffer() local 92 const TfLiteTensor* scales, const TfLiteTensor* embeddings, in TFLiteEmbeddingExecutor()
|
/external/pytorch/aten/src/ATen/native/quantized/ |
D | AffineQuantizer.cpp | 95 const Tensor& scales, in checkPerChannelParamsSize() 141 const Tensor& scales, in quantize_tensor_per_channel_affine() 176 const Tensor& scales, in quantize_tensor_per_channel_float_qparams() 233 const Tensor& scales, in dequantize_tensor_per_channel_affine() 267 const Tensor& scales, in dequantize_tensor_per_channel_float_qparams()
|
/external/tensorflow/tensorflow/core/kernels/uniform_quant_ops/ |
D | uniform_dequantize_op.cc | 32 void EvalPerChannelDequantize(const Tensor& input, const Tensor& scales, in EvalPerChannelDequantize() 51 void EvalDequantize(const Tensor& input, const Tensor& scales, in EvalDequantize() 82 const Tensor& scales = context->input(1); in Compute() local
|
D | uniform_quantize_op.cc | 34 void EvalPerChannelQuantize(const Tensor& input, const Tensor& scales, in EvalPerChannelQuantize() 56 void EvalQuantize(const Tensor& input, const Tensor& scales, in EvalQuantize() 98 const Tensor& scales = context->input(1); in Compute() local
|
/external/pytorch/aten/src/ATen/benchmarks/ |
D | quantize_per_channel.cpp | 13 at::Tensor scales = at::rand({channels}); in quantize_per_channel_4d_contiguous() local 33 at::Tensor scales = at::rand({channels}); in quantize_per_channel_4d_channels_last() local 49 at::Tensor scales = at::rand({channels}); in quantize_per_channel_2d() local
|
/external/pytorch/aten/src/ATen/native/cpu/ |
D | int8mm_kernel.cpp | 35 const BFloat16* RESTRICT scales, in tinygemm_kernel() 230 const T* RESTRICT scales, in tinygemm_kernel_() 272 const Half* RESTRICT scales, in tinygemm_kernel() 285 const BFloat16* RESTRICT scales, in tinygemm_kernel() 298 const float* RESTRICT scales, in tinygemm_kernel() 313 const T* RESTRICT scales, in tinygemm_kernel() 363 const Tensor& scales) { in int8pack_mm_kernel_() 424 const Tensor& scales) { in int8pack_mm_kernel()
|
/external/pytorch/aten/src/ATen/native/quantized/cuda/ |
D | AffineQuantizer.cu | 90 const Tensor& scales, in quantize_tensor_per_channel_affine_cuda() 133 const Tensor& scales, in dequantize_tensor_per_channel_affine_cuda() 169 const Tensor& scales, in quantize_tensor_per_channel_float_qparams_cuda() 217 const Tensor& scales, in dequantize_tensor_per_channel_float_qparams_cuda()
|
D | MakePerTensorQuantizedTensor.cu | 51 const Tensor& scales, in make_per_channel_quantized_tensor_cuda()
|
/external/pytorch/aten/src/ATen/native/quantized/cpu/ |
D | LinearUnpackImpl.cpp | 36 auto scales = at::from_blob( in unpack() local 83 auto scales = at::from_blob( in unpack() local
|
D | qconv_unpack_impl.cpp | 60 auto scales = at::from_blob( in unpack() local
|
/external/executorch/backends/vulkan/runtime/graph/ops/impl/ |
D | Upsample.cpp | 34 DoubleListPtr scales = graph->get_double_list(scale_factors); in resize_upsample_nearest2d_node() local 85 auto scales = graph.get_double_list(scale_factors); in add_upsample_nearest2d_node() local
|
D | QuantizedLinear.cpp | 22 const ValueRef scales, in check_q_8w_linear_args() 86 ValueRef scales = prepack_standard( in add_q_8w_linear_node() local 157 ValueRef scales = in add_q_8w_linear_optimized_node() local
|
/external/pytorch/aten/src/ATen/native/ao_sparse/quantized/cpu/ |
D | qlinear_unpack.cpp | 35 at::Tensor scales = at::empty( in unpack() local 82 at::Tensor scales = at::empty( in unpack() local
|
/external/pytorch/aten/src/ATen/native/mkldnn/ |
D | BinaryOps.cpp | 91 const std::vector<float> scales{alpha.to<float>(), 1.0}; in mkldnn_add_out() local 94 const std::vector<float> scales{1.0, alpha.to<float>()}; in mkldnn_add_out() local 110 const std::vector<float> scales{1.0, alpha.to<float>()}; in mkldnn_add() local
|
/external/executorch/extension/llm/custom_ops/ |
D | test_update_quantized_cache.py | 46 def _update_k(self, start_pos, value, scales, zero_points): argument 52 def _update_v(self, start_pos, value, scales, zero_points): argument
|
/external/tensorflow/tensorflow/lite/tools/optimize/ |
D | quantization_utils.cc | 204 std::vector<float>* scales, in GetSymmetricScalesFromMaxMin() 373 std::vector<float> scales(1, scaling_factor); in SymmetricQuantizeFloatsToInt16() local 544 TfLiteStatus AddQuantizationParams(const std::vector<float>& scales, in AddQuantizationParams() 595 std::vector<float> scales(channel_dim_size); in SymmetricQuantizeTensorPerChannel() local 615 const std::vector<float>& scales) { in SymmetricBiasQuantize() 649 std::vector<float> scales(1, scaling_factor); in SymmetricPerLayerBiasQuantize() local 675 std::vector<float> scales(number_of_dimension); in SymmetricPerChannelBiasQuantize() local
|
/external/executorch/backends/qualcomm/aot/wrappers/ |
D | QuantizeParamsWrapper.cpp | 33 std::vector<float> scales( in CreateQuantizationParamWrapper() local
|
/external/executorch/examples/models/llama/experimental/ |
D | subclass.py | 143 scales, argument 153 scales, argument
|
/external/tensorflow/tensorflow/lite/testing/op_tests/ |
D | nearest_upsample.py | 33 def new_shape_for_upsample(original_shape, scales, axis): argument
|
/external/pytorch/test/quantization/core/ |
D | test_quantized_tensor.py | 604 def _test_quantize_per_channel(self, r, scales, zero_points, axis, float_params): argument 606 def _quantize_per_channel_ref_nd(data, scales, zero_points, float_params): argument 639 def quantize_c(data, scales, zero_points): argument 678 def quantize_ref(data, scales, zero_points): argument 726 def _quantize_per_channel_sub_byte_ref(data, scales, zero_points, axis, bit_width): argument
|
/external/pytorch/aten/src/ATen/test/ |
D | quantized_test.cpp | 126 auto scales = rand({numel}).toType(kDouble); in TEST() local 154 auto scales = rand({C}).toType(kDouble); in TEST() local 183 auto scales = rand({C}).toType(kDouble); in TEST() local 259 auto scales = rand({C}).toType(kDouble); in TEST() local
|