Home
last modified time | relevance | path

Searched defs:scales (Results 1 – 25 of 129) sorted by relevance

123456

/external/tensorflow/tensorflow/lite/kernels/internal/
Dper_channel_dequantize_test.cc30 const std::vector<float> scales = {0.5, 0.25}; in TEST() local
52 const std::vector<float> scales = {0.5, 0.25, 0.5, 0.25, 1.0}; in TEST() local
74 const std::vector<float> scales = {0.5, 0.25}; in TEST() local
98 const std::vector<float> scales = {0.5, 0.25, 0.5, 0.25, 1.0}; in TEST() local
Dper_channel_quantize_test.cc28 const std::vector<float> scales = {0.5, 0.25}; in TEST() local
49 const std::vector<float> scales = {0.5, 0.25, 0.5, 0.25, 1.0}; in TEST() local
70 const std::vector<float> scales = {0.5, 0.25}; in TEST() local
94 const std::vector<float> scales = {0.5, 0.25, 0.5, 0.25, 1.0}; in TEST() local
/external/libtextclassifier/native/annotator/
Dquantization_test.cc40 std::vector<float> scales{{0.1, 9.0, -7.0}}; in TEST() local
92 std::vector<float> scales(num_buckets); in TEST() local
115 std::vector<float> scales(num_buckets, 1.0); in TEST() local
135 std::vector<float> scales(num_buckets, 1.0); in TEST() local
Dquantization.cc28 void DequantizeAdd8bit(const float* scales, const uint8* embeddings, in DequantizeAdd8bit()
40 void DequantizeAddNBit(const float* scales, const uint8* embeddings, in DequantizeAddNBit()
73 bool DequantizeAdd(const float* scales, const uint8* embeddings, in DequantizeAdd()
Dmodel-executor.cc71 const TfLiteTensor* scales = interpreter->tensor(1); in FromBuffer() local
92 const TfLiteTensor* scales, const TfLiteTensor* embeddings, in TFLiteEmbeddingExecutor()
/external/pytorch/aten/src/ATen/native/quantized/
DAffineQuantizer.cpp95 const Tensor& scales, in checkPerChannelParamsSize()
141 const Tensor& scales, in quantize_tensor_per_channel_affine()
176 const Tensor& scales, in quantize_tensor_per_channel_float_qparams()
233 const Tensor& scales, in dequantize_tensor_per_channel_affine()
267 const Tensor& scales, in dequantize_tensor_per_channel_float_qparams()
/external/tensorflow/tensorflow/core/kernels/uniform_quant_ops/
Duniform_dequantize_op.cc32 void EvalPerChannelDequantize(const Tensor& input, const Tensor& scales, in EvalPerChannelDequantize()
51 void EvalDequantize(const Tensor& input, const Tensor& scales, in EvalDequantize()
82 const Tensor& scales = context->input(1); in Compute() local
Duniform_quantize_op.cc34 void EvalPerChannelQuantize(const Tensor& input, const Tensor& scales, in EvalPerChannelQuantize()
56 void EvalQuantize(const Tensor& input, const Tensor& scales, in EvalQuantize()
98 const Tensor& scales = context->input(1); in Compute() local
/external/pytorch/aten/src/ATen/benchmarks/
Dquantize_per_channel.cpp13 at::Tensor scales = at::rand({channels}); in quantize_per_channel_4d_contiguous() local
33 at::Tensor scales = at::rand({channels}); in quantize_per_channel_4d_channels_last() local
49 at::Tensor scales = at::rand({channels}); in quantize_per_channel_2d() local
/external/pytorch/aten/src/ATen/native/cpu/
Dint8mm_kernel.cpp35 const BFloat16* RESTRICT scales, in tinygemm_kernel()
230 const T* RESTRICT scales, in tinygemm_kernel_()
272 const Half* RESTRICT scales, in tinygemm_kernel()
285 const BFloat16* RESTRICT scales, in tinygemm_kernel()
298 const float* RESTRICT scales, in tinygemm_kernel()
313 const T* RESTRICT scales, in tinygemm_kernel()
363 const Tensor& scales) { in int8pack_mm_kernel_()
424 const Tensor& scales) { in int8pack_mm_kernel()
/external/pytorch/aten/src/ATen/native/quantized/cuda/
DAffineQuantizer.cu90 const Tensor& scales, in quantize_tensor_per_channel_affine_cuda()
133 const Tensor& scales, in dequantize_tensor_per_channel_affine_cuda()
169 const Tensor& scales, in quantize_tensor_per_channel_float_qparams_cuda()
217 const Tensor& scales, in dequantize_tensor_per_channel_float_qparams_cuda()
DMakePerTensorQuantizedTensor.cu51 const Tensor& scales, in make_per_channel_quantized_tensor_cuda()
/external/pytorch/aten/src/ATen/native/quantized/cpu/
DLinearUnpackImpl.cpp36 auto scales = at::from_blob( in unpack() local
83 auto scales = at::from_blob( in unpack() local
Dqconv_unpack_impl.cpp60 auto scales = at::from_blob( in unpack() local
/external/executorch/backends/vulkan/runtime/graph/ops/impl/
DUpsample.cpp34 DoubleListPtr scales = graph->get_double_list(scale_factors); in resize_upsample_nearest2d_node() local
85 auto scales = graph.get_double_list(scale_factors); in add_upsample_nearest2d_node() local
DQuantizedLinear.cpp22 const ValueRef scales, in check_q_8w_linear_args()
86 ValueRef scales = prepack_standard( in add_q_8w_linear_node() local
157 ValueRef scales = in add_q_8w_linear_optimized_node() local
/external/pytorch/aten/src/ATen/native/ao_sparse/quantized/cpu/
Dqlinear_unpack.cpp35 at::Tensor scales = at::empty( in unpack() local
82 at::Tensor scales = at::empty( in unpack() local
/external/pytorch/aten/src/ATen/native/mkldnn/
DBinaryOps.cpp91 const std::vector<float> scales{alpha.to<float>(), 1.0}; in mkldnn_add_out() local
94 const std::vector<float> scales{1.0, alpha.to<float>()}; in mkldnn_add_out() local
110 const std::vector<float> scales{1.0, alpha.to<float>()}; in mkldnn_add() local
/external/executorch/extension/llm/custom_ops/
Dtest_update_quantized_cache.py46 def _update_k(self, start_pos, value, scales, zero_points): argument
52 def _update_v(self, start_pos, value, scales, zero_points): argument
/external/tensorflow/tensorflow/lite/tools/optimize/
Dquantization_utils.cc204 std::vector<float>* scales, in GetSymmetricScalesFromMaxMin()
373 std::vector<float> scales(1, scaling_factor); in SymmetricQuantizeFloatsToInt16() local
544 TfLiteStatus AddQuantizationParams(const std::vector<float>& scales, in AddQuantizationParams()
595 std::vector<float> scales(channel_dim_size); in SymmetricQuantizeTensorPerChannel() local
615 const std::vector<float>& scales) { in SymmetricBiasQuantize()
649 std::vector<float> scales(1, scaling_factor); in SymmetricPerLayerBiasQuantize() local
675 std::vector<float> scales(number_of_dimension); in SymmetricPerChannelBiasQuantize() local
/external/executorch/backends/qualcomm/aot/wrappers/
DQuantizeParamsWrapper.cpp33 std::vector<float> scales( in CreateQuantizationParamWrapper() local
/external/executorch/examples/models/llama/experimental/
Dsubclass.py143 scales, argument
153 scales, argument
/external/tensorflow/tensorflow/lite/testing/op_tests/
Dnearest_upsample.py33 def new_shape_for_upsample(original_shape, scales, axis): argument
/external/pytorch/test/quantization/core/
Dtest_quantized_tensor.py604 def _test_quantize_per_channel(self, r, scales, zero_points, axis, float_params): argument
606 def _quantize_per_channel_ref_nd(data, scales, zero_points, float_params): argument
639 def quantize_c(data, scales, zero_points): argument
678 def quantize_ref(data, scales, zero_points): argument
726 def _quantize_per_channel_sub_byte_ref(data, scales, zero_points, axis, bit_width): argument
/external/pytorch/aten/src/ATen/test/
Dquantized_test.cpp126 auto scales = rand({numel}).toType(kDouble); in TEST() local
154 auto scales = rand({C}).toType(kDouble); in TEST() local
183 auto scales = rand({C}).toType(kDouble); in TEST() local
259 auto scales = rand({C}).toType(kDouble); in TEST() local

123456