Home
last modified time | relevance | path

Searched refs:proj_clip (Results 1 – 25 of 40) sorted by relevance

12

/external/tensorflow/tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/
Dbasic_lstm.mlir6 …0e+00 : f32, fused_activation_function = "RELU", kernel_type = "BASIC", proj_clip = 2.000000e+00 :…
8 …rg3, %arg4) {fused_activation_function = "RELU", cell_clip = 1.0 : f32, proj_clip = 2.0 : f32} : (…
Dlstm.mlir7 …00e+00 : f32, fused_activation_function = "NONE", kernel_type = "FULL", proj_clip = 0.000000e+00 :…
14 …00e+00 : f32, fused_activation_function = "NONE", kernel_type = "FULL", proj_clip = 0.000000e+00 :…
23 …!quant.uniform<i8:f32, 0.0075630000792443752:2>>, kernel_type = "FULL", proj_clip = 0.01 : f32} : …
28 …x!quant.uniform<i16:f32, 0.0057529998011887074>>, kernel_type = "FULL", proj_clip = 0.00999999977 …
35 …rmediate = tensor<0xf32>, input_to_output_intermediate = tensor<0xf32>, proj_clip = 0.000000e+00 :…
36 …rmediate = tensor<0xf32>, input_to_output_intermediate = tensor<0xf32>, proj_clip = 0.000000e+00 :…
Dtest_schema.fbs613 proj_clip: float; // Optional, 0.0 means no clipping
627 proj_clip: float; // Optional, 0.0 means no clipping
640 proj_clip: float; // Optional, 0.0 means no clipping
/external/tensorflow/tensorflow/lite/c/
Dbuiltin_op_data.h267 float proj_clip; member
281 float proj_clip; member
295 float proj_clip; member
/external/tensorflow/tensorflow/lite/tools/optimize/calibration/builtin_logging_ops/
Dlstm.cc127 const float proj_clip, float* output_state, float* scratch, Logger* logger, in CalculateLstmOutputCalibration() argument
149 if (proj_clip > 0.0f) { in CalculateLstmOutputCalibration()
150 tensor_utils::CwiseClipping(output_state, n_batch * n_output, proj_clip); in CalculateLstmOutputCalibration()
253 params->proj_clip, output_state_ptr, scratch2, logger, in LstmStepCalibration()
595 lstm_params.proj_clip = params->proj_clip; in lstm_eval()
/external/tensorflow/tensorflow/lite/kernels/
Doptional_tensor_test.cc33 bool use_projection_bias, float cell_clip, float proj_clip, in LSTMOpModel() argument
106 cell_clip, proj_clip) in LSTMOpModel()
Dunidirectional_sequence_lstm.cc55 const float proj_clip = params->proj_clip; in PopulateQuantizedLstmParams8x8_16() local
76 if (proj_clip > 0.0) { in PopulateQuantizedLstmParams8x8_16()
78 std::max(proj_clip / proj_params->scale->data[0], -128.0f), 127.0f)); in PopulateQuantizedLstmParams8x8_16()
419 TF_LITE_ENSURE(context, params->proj_clip >= 0); in CheckInputTensorDimensions()
1311 lstm_params.proj_clip = params->proj_clip; in Eval()
Dunidirectional_sequence_lstm_test.cc36 float proj_clip, in UnidirectionalLSTMOpModel() argument
131 proj_clip, time_major, asymmetric_quantize_inputs) in UnidirectionalLSTMOpModel()
291 float proj_clip, const std::vector<std::vector<int>>& input_shapes, in HybridUnidirectionalLSTMOpModel() argument
296 cell_clip, proj_clip, input_shapes, tensor_type, false, in HybridUnidirectionalLSTMOpModel()
2473 float proj_clip, const std::vector<std::vector<int>>& input_shapes, in LayerNormUnidirectionalLSTMOpModel() argument
2478 cell_clip, proj_clip, input_shapes, TensorType_FLOAT32, true) {} in LayerNormUnidirectionalLSTMOpModel()
Dlstm_eval.cc267 const float proj_clip, float* output_state, in CalculateLstmOutputFloat() argument
286 if (proj_clip > 0.0f) { in CalculateLstmOutputFloat()
287 tensor_utils::CwiseClipping(output_state, n_batch * n_output, proj_clip); in CalculateLstmOutputFloat()
437 const float proj_clip, float* output_state, bool asymmetric_quantize_inputs, in CalculateLstmOutputHybrid() argument
477 if (proj_clip > 0.0f) { in CalculateLstmOutputHybrid()
478 tensor_utils::CwiseClipping(output_state, n_batch * n_output, proj_clip); in CalculateLstmOutputHybrid()
898 params->proj_clip, output_state_ptr, scratch2); in LstmStepFloat()
1212 projection_weights_scale, projection_bias_ptr, params->proj_clip, in LstmStepHybrid()
Dlstm.cc147 const float proj_clip = params->proj_clip; in PopulateQuantizedLstmParams8x8_16() local
167 if (proj_clip > 0.0) { in PopulateQuantizedLstmParams8x8_16()
169 std::max(proj_clip / proj_params->scale->data[0], -128.0f), 127.0f)); in PopulateQuantizedLstmParams8x8_16()
809 const float proj_clip = params->proj_clip; in PopulateQuantizedLstmParams8x8_8() local
827 if (proj_clip > 0.0) { in PopulateQuantizedLstmParams8x8_8()
829 std::max(proj_clip / proj_params->scale->data[0], -128.0f), 127.0f)); in PopulateQuantizedLstmParams8x8_8()
862 TF_LITE_ENSURE(context, params->proj_clip >= 0); in CheckInputTensorDimensions()
/external/tensorflow/tensorflow/compiler/mlir/lite/tests/mlir2flatbuffer/
Dbasic_lstm.mlir85 // CHECK-NEXT: proj_clip: 2.0,
124 …rg3, %arg4) {fused_activation_function = "RELU", cell_clip = 1.0 : f32, proj_clip = 2.0 : f32} : (…
Dlstm_quantized.mlir5 …form<i8<-127:127>:f32, 0.0075630000792443752:2>>, kernel_type = "FULL", proj_clip = 0.01 : f32} : …
263 // CHECK-NEXT: proj_clip: 0.01
Dlstm.mlir269 …00e+00 : f32, fused_activation_function = "NONE", kernel_type = "FULL", proj_clip = 0.000000e+00 :…
/external/tensorflow/tensorflow/lite/delegates/gpu/common/
Dlstm_parser.cc340 bool has_projection, float proj_clip, in BuildOutputStateUpdate() argument
395 if (proj_clip <= 0.0f) { in BuildOutputStateUpdate()
407 attr.param = proj_clip; in BuildOutputStateUpdate()
418 attr.param = -proj_clip; in BuildOutputStateUpdate()
530 params->proj_clip, &new_output_state)); in ParseLSTMAttributes()
/external/tensorflow/tensorflow/lite/experimental/examples/lstm/
Drnn_cell.py203 proj_clip=None, argument
271 self._proj_clip = proj_clip
/external/tensorflow/tensorflow/lite/core/api/
Dflatbuffer_conversions.cc568 params->proj_clip = lstm_params->proj_clip(); in ParseOpDataTfLite()
601 params->proj_clip = seq_lstm_params->proj_clip(); in ParseOpDataTfLite()
618 params->proj_clip = bidi_lstm_params->proj_clip(); in ParseOpDataTfLite()
/external/tensorflow/tensorflow/compiler/mlir/lite/tests/
Dprepare-quantize-post-training.mlir21 …2) {cell_clip = 1.000000e+01 : f32, fused_activation_function = "TANH", proj_clip = 0.000000e+00 :…
79 proj_clip = 0.000000e+00 : f32, time_major = false} : (
157 proj_clip = 0.000000e+00 : f32,time_major = false} : (
243 proj_clip = 0.000000e+00 : f32, time_major = false} : (
335 proj_clip = 0.000000e+00 : f32,time_major = false} : (
Dprepare-composite-functions-tf.mlir89 …00e+01 : f32, fused_activation_function = "TANH", kernel_type = "FULL", proj_clip = 0.000000e+00 :…
149 …00e+01 : f32, fused_activation_function = "TANH", kernel_type = "FULL", proj_clip = 0.000000e+00 :…
207 …]) {cell_clip = 1.000000e+01 : f32, fused_activation_function = "TANH", proj_clip = 0.000000e+00 :…
248 …]) {cell_clip = 1.000000e+01 : f32, fused_activation_function = "TANH", proj_clip = 0.000000e+00 :…
292 …]) {cell_clip = 1.000000e+01 : f32, fused_activation_function = "TANH", proj_clip = 0.000000e+00 :…
336 …]) {cell_clip = 1.000000e+01 : f32, fused_activation_function = "TANH", proj_clip = 0.000000e+00 :…
385 …]) {cell_clip = 1.000000e+01 : f32, fused_activation_function = "TANH", proj_clip = 0.000000e+00 :…
435 …]) {cell_clip = 1.000000e+01 : f32, fused_activation_function = "TANH", proj_clip = 0.000000e+00 :…
/external/tensorflow/tensorflow/lite/tools/optimize/calibration/custom_logging_ops/
Dlstm.cc314 if (params->proj_clip > 0.0) { in LstmStepWithAuxInput()
316 n_output, params->proj_clip); in LstmStepWithAuxInput()
/external/tensorflow/tensorflow/lite/schema/
Dschema_v0.fbs196 proj_clip: float; // Optional, 0.0 means no clipping
Dschema_v2.fbs217 proj_clip: float; // Optional, 0.0 means no clipping
Dschema_v1.fbs209 proj_clip: float; // Optional, 0.0 means no clipping
Dschema_v3a.fbs619 proj_clip: float; // Optional, 0.0 means no clipping
633 proj_clip: float; // Optional, 0.0 means no clipping
646 proj_clip: float; // Optional, 0.0 means no clipping
/external/tensorflow/tensorflow/tools/api/golden/v1/
Dtensorflow.nn.rnn_cell.-l-s-t-m-cell.pbtxt151 …s\', \'use_peepholes\', \'cell_clip\', \'initializer\', \'num_proj\', \'proj_clip\', \'num_unit_sh…
Dtensorflow.lite.experimental.nn.-t-f-lite-l-s-t-m-cell.pbtxt151 …s\', \'use_peepholes\', \'cell_clip\', \'initializer\', \'num_proj\', \'proj_clip\', \'num_unit_sh…

12