| /external/tensorflow/tensorflow/lite/kernels/internal/reference/ |
| D | portable_tensor_utils.h | 60 int n_batch, float* result) { in MatrixBatchVectorMultiplyAccumulate() 69 int n_batch, in MatrixBatchVectorMultiplyAccumulate() 78 int n_batch, float* __restrict__ result, const float* per_channel_scale, in MatrixBatchVectorMultiplyAccumulate() 91 int n_batch, int32_t* scratch, in MatrixBatchVectorMultiplyAccumulate() 101 const float* __restrict__ vector, int n_batch, float* __restrict__ result) { in SparseMatrixBatchVectorMultiplyAccumulate1x4() 108 int m_rows, int m_cols, const float* __restrict__ vector, int n_batch, in SparseMatrixBatchVectorMultiplyAccumulate() 117 const float* scaling_factors, int n_batch, float* __restrict__ result) { in SparseMatrixBatchVectorMultiplyAccumulate() 126 int32_t n_batch, int32_t n_input, int32_t n_output, int32_t output_zp, in MatrixBatchVectorMultiplyAccumulate() 136 int32_t n_batch, int32_t n_input, int32_t n_output, int32_t output_zp, in MatrixBatchVectorMultiplyAccumulate() 153 int32_t n_batch, int32_t n_input, int32_t n_cell, in MatrixBatchVectorMultiply() [all …]
|
| D | portable_tensor_utils.cc | 122 int n_batch, float* result) { in PortableMatrixBatchVectorMultiplyAccumulate() 141 int n_batch, float* __restrict__ result) { in PortableMatrixBatchVectorMultiplyAccumulate() 166 int n_batch, float* __restrict__ result, const float* per_channel_scale, in PortableMatrixBatchVectorMultiplyAccumulate() 209 const float* __restrict__ vector, int n_batch, float* __restrict__ result) { in PortableSparseMatrixBatchVectorMultiplyAccumulate1x4() 232 int m_rows, int m_cols, const float* __restrict__ vector, int n_batch, in PortableSparseMatrixBatchVectorMultiplyAccumulate() 262 const float* scaling_factors, int n_batch, float* __restrict__ result) { in PortableSparseMatrixBatchVectorMultiplyAccumulate() 296 int32_t n_batch, int32_t n_input, int32_t n_output, int32_t output_zp, in PortableMatrixBatchVectorMultiplyAccumulateImpl() 325 int32_t n_batch, int32_t n_input, int32_t n_output, int32_t output_zp, in PortableMatrixBatchVectorMultiplyAccumulate() 335 int32_t n_batch, int32_t n_input, int32_t n_output, int32_t output_zp, in PortableMatrixBatchVectorMultiplyAccumulate() 347 int32_t n_batch, int32_t n_input, in PortableMatrixBatchVectorMultiply() [all …]
|
| D | svdf.h | 77 const int n_batch = input_shape.Dims(0); in EvalIntegerSVDF() local
|
| /external/tensorflow/tensorflow/lite/kernels/internal/optimized/ |
| D | sse_tensor_utils.h | 39 int n_batch, float* result) { in MatrixBatchVectorMultiplyAccumulate() 47 const float* __restrict__ scaling_factors, int n_batch, in MatrixBatchVectorMultiplyAccumulate() 56 int n_batch, float* __restrict__ result, const float* per_channel_scale, in MatrixBatchVectorMultiplyAccumulate() 67 const float* __restrict__ scaling_factors, int n_batch, in MatrixBatchVectorMultiplyAccumulate() 77 const float* __restrict__ vector, int n_batch, float* __restrict__ result) { in SparseMatrixBatchVectorMultiplyAccumulate1x4() 84 int m_rows, int m_cols, const float* __restrict__ vector, int n_batch, in SparseMatrixBatchVectorMultiplyAccumulate() 93 const float* __restrict__ scaling_factors, int n_batch, in SparseMatrixBatchVectorMultiplyAccumulate() 102 int32_t n_batch, int32_t n_input, int32_t n_output, int32_t output_zp, in MatrixBatchVectorMultiplyAccumulate() 112 int32_t n_batch, int32_t n_input, int32_t n_output, int32_t output_zp, in MatrixBatchVectorMultiplyAccumulate() 123 int32_t n_batch, int32_t n_input, int32_t n_cell, in MatrixBatchVectorMultiply() [all …]
|
| D | neon_tensor_utils.h | 29 int n_batch, float* result) { in MatrixBatchVectorMultiplyAccumulate() 38 int n_batch, in MatrixBatchVectorMultiplyAccumulate() 48 int n_batch, int32_t* scratch, in MatrixBatchVectorMultiplyAccumulate() 58 int n_batch, float* __restrict__ result, const float* per_channel_scale, in MatrixBatchVectorMultiplyAccumulate() 69 const float* __restrict__ vector, int n_batch, float* __restrict__ result) { in SparseMatrixBatchVectorMultiplyAccumulate1x4() 76 int m_rows, int m_cols, const float* __restrict__ vector, int n_batch, in SparseMatrixBatchVectorMultiplyAccumulate() 85 const float* scaling_factors, int n_batch, float* __restrict__ result) { in SparseMatrixBatchVectorMultiplyAccumulate() 93 int32_t n_batch, int32_t n_input, int32_t n_output, int32_t output_zp, in MatrixBatchVectorMultiplyAccumulate() 103 int32_t n_batch, int32_t n_input, int32_t n_output, int32_t output_zp, in MatrixBatchVectorMultiplyAccumulate() 114 int32_t n_batch, int32_t n_input, int32_t n_cell, in MatrixBatchVectorMultiply() [all …]
|
| D | neon_tensor_utils.cc | 202 int n_batch, float* result) { in NeonMatrixBatchVectorMultiplyAccumulate() 251 const int8_t* ShuffleVectors(const int8_t* vectors, const int n_batch, in ShuffleVectors() 306 const int8_t* vectors, const float* scaling_factors, int n_batch, in DotprodMatrixBatchFourVectorMultiplyAccumulate() 434 const int8_t* vectors, const float* scaling_factors, int n_batch, in DotprodMatrixBatchFourVectorMultiplyAccumulate() 563 const int8_t* vectors, const float* scaling_factors, int n_batch, in DotprodMatrixBatchFourVectorMultiplyAccumulate() 596 const int8_t* vectors, const float* scaling_factors, int n_batch, in DotprodMatrixBatchPaddedFourVectorMultiplyAccumulate() 668 const int8_t* vectors, const float* scaling_factors, int n_batch, in DotprodMatrixBatchPaddedFourVectorMultiplyAccumulate() 679 const float* scaling_factors, int n_batch, float* __restrict__ result) { in DotprodSparseMatrixBatchVectorMultiplyAccumulate() 743 int32_t n_batch, int32_t n_input, in NeonMatrixBatchVectorMultiplyImpl() 853 int32_t multiplier, int32_t shift, int32_t n_batch, int32_t n_output, in NeonMatrixBatchVectorAccumulateImpl() [all …]
|
| D | sse_tensor_utils.cc | 99 const float* __restrict__ scaling_factors, int n_batch, in SseMatrixBatchVectorMultiplyAccumulateImpl() 177 const int8_t* input_to_gate_weights, int32_t n_batch, in SseCpuBackendGemm() 211 const float* __restrict__ scaling_factors, int n_batch, in SseMatrixBatchVectorMultiplyAccumulate() 222 const float* __restrict__ scaling_factors, int n_batch, int32_t* scratch, in SseMatrixBatchVectorMultiplyAccumulate() 272 const float* __restrict__ scaling_factors, int n_batch, in SseMatrixBatchVectorMultiplyAccumulate() 394 const float* __restrict__ scaling_factors, int n_batch, in SseSparseMatrixBatchVectorMultiplyAccumulate()
|
| /external/tensorflow/tensorflow/lite/kernels/internal/ |
| D | tensor_utils_common.h | 63 inline void BatchQuantizeFloats(const float* float_data_ptr, int n_batch, in BatchQuantizeFloats() 352 int v_size, int n_batch, in BatchVectorBatchVectorDotProduct() 369 const T* batch_vector, int n_batch, in VectorBatchVectorCwiseProduct() 385 int n_batch, T* result) { in VectorBatchVectorCwiseProductAccumulate() 402 void VectorBatchVectorAdd(const T* vector, int v_size, int n_batch, in VectorBatchVectorAdd() 414 void VectorBatchVectorAssign(const T* vector, int v_size, int n_batch, in VectorBatchVectorAssign()
|
| D | tensor_utils.h | 63 const float* vector_scaling_factors, int n_batch, in MatrixBatchVectorMultiplyAccumulate()
|
| /external/tensorflow/tensorflow/lite/delegates/gpu/cl/kernels/ |
| D | lstm_full_test.cc | 35 LSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, bool use_cifg, in LSTMOpModel() 265 const int n_batch = 1; in TEST_P() local 339 const int n_batch = 1; in TEST_P() local 407 const int n_batch = 1; in TEST_P() local 970 const int n_batch = 1; in TEST_F() local 1056 const int n_batch = 1; in TEST_F() local 1120 const int n_batch = 1; in TEST_F() local
|
| /external/tensorflow/tensorflow/lite/kernels/ |
| D | unidirectional_sequence_lstm_test.cc | 32 UnidirectionalLSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, in UnidirectionalLSTMOpModel() 288 int n_batch, int n_input, int n_cell, int n_output, int sequence_length, in HybridUnidirectionalLSTMOpModel() 500 const int n_batch = 1; in TEST_F() local 562 const int n_batch = 1; in TEST_F() local 629 const int n_batch = 1; in TEST_P() local 692 const int n_batch = 1; in TEST_P() local 803 const int n_batch = 1; in TEST_F() local 865 const int n_batch = 1; in TEST_P() local 927 const int n_batch = 1; in TEST_P() local 1590 const int n_batch = 2; in TEST_F() local [all …]
|
| D | bidirectional_sequence_lstm_test.cc | 32 BidirectionalLSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, in BidirectionalLSTMOpModel() 430 const int n_batch = 1; in TEST_P() local 596 const int n_batch = 2; in TEST_P() local 762 const int n_batch = 1; in TEST() local 923 const int n_batch = 1; in TEST() local 1074 const int n_batch = 1; in TEST() local 1226 const int n_batch = 2; in TEST() local 1929 const int n_batch = 2; in TEST() local 2630 const int n_batch = 1; in TEST_P() local 2804 const int n_batch = 1; in TEST_P() local
|
| D | lstm_test.cc | 39 LSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, bool use_cifg, in LSTMOpModel() 413 const int n_batch = 1; in TEST_P() local 490 const int n_batch = 1; in TEST_P() local 561 const int n_batch = 2; in TEST_P() local 1144 const int n_batch = 2; in TEST_P() local 1239 const int n_batch = 2; in TEST_P() local 1314 LSTMIntegerOpModel(int n_batch, int n_input, int n_cell, int n_output, in LSTMIntegerOpModel() 1598 const int n_batch = 2; in TEST() local 1760 const int n_batch = 2; in TEST() local 1933 const int n_batch = 2; in TEST() local [all …]
|
| D | lstm_eval.cc | 156 const int n_batch, const int n_input, const int n_aux_input, in CalculateLstmGateFloat() 218 void UpdateLstmCellFloat(int n_batch, int n_cell, float* cell_state, in UpdateLstmCellFloat() 262 void CalculateLstmOutputFloat(int n_batch, int n_cell, int n_output, in CalculateLstmOutputFloat() 321 const int n_batch, const int n_input, const int n_aux_input, in CalculateLstmGateHybrid() 433 int n_batch, int n_cell, int n_output, const float* cell_state, in CalculateLstmOutputHybrid() 506 const int n_batch, const int n_input, const int n_output, const int n_cell, in CalculateLstmGateInteger8x8_16() 571 void UpdateLstmCellInteger(int n_batch, int n_cell, int16_t* cell_state, in UpdateLstmCellInteger() 617 int n_batch, int n_cell, int n_output, const int16_t* cell_state, in CalculateLstmOutputInteger8x8_16() 671 const int n_batch, const int n_input, const int n_output, const int n_cell, in CalculateLstmGateInteger8x8_8() 725 int n_batch, int n_cell, int n_output, const int16_t* cell_state, in CalculateLstmOutputInteger8x8_8() [all …]
|
| D | optional_tensor_test.cc | 31 LSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, bool use_cifg, in LSTMOpModel() 229 const int n_batch = 1; in TEST() local
|
| D | unidirectional_sequence_lstm.cc | 885 const int n_batch = time_major ? input->dims->data[1] : input->dims->data[0]; in Prepare() local
|
| /external/tensorflow/tensorflow/lite/experimental/kernels/ |
| D | unidirectional_sequence_gru_test.cc | 34 explicit GRUOpModel(int n_batch, int n_input, int n_output, in GRUOpModel() 101 const int n_batch = 2; in TEST() local
|
| D | unidirectional_sequence_gru.cc | 38 const int n_batch = input->dims->data[1]; in GruImpl() local 134 const int n_batch = input->dims->data[1]; in Prepare() local
|
| D | gru_cell.cc | 45 const int n_batch = input_shape.Dims(0); in GruCell() local
|
| /external/tensorflow/tensorflow/lite/tools/optimize/calibration/builtin_logging_ops/ |
| D | lstm.cc | 46 const int n_batch, const int n_input, const int n_aux_input, in CalculateLstmGateFloat() 100 void UpdateLstmCellFloat(int n_batch, int n_cell, float* cell_state, in UpdateLstmCellFloat() 124 int n_batch, int n_cell, int n_output, const float* cell_state, in CalculateLstmOutputCalibration() 180 const TfLiteLSTMParams* params, int n_batch, int n_cell, int n_input, in LstmStepCalibration() 293 int max_time, n_batch; in EvalCalibration() local
|
| /external/tensorflow/tensorflow/lite/micro/kernels/xtensa/ |
| D | svdf.cc | 63 const int n_batch = input_tensor->dims->data[0]; in EvalIntegerSvdfHifimini() local 258 const int n_batch = input_tensor->dims->data[0]; in EvalIntegerSvdfHifi4() local
|
| /external/tensorflow/tensorflow/lite/micro/kernels/ |
| D | svdf_common.cc | 42 const int n_batch = input_tensor->dims->data[0]; in EvalIntegerSvdfReference() local
|
| /external/tensorflow/tensorflow/lite/tools/optimize/calibration/custom_logging_ops/ |
| D | lstm.cc | 63 const TfLiteLSTMParams* params, int n_batch, int n_cell, int n_input, in LstmStepWithAuxInput() 362 int max_time, n_batch; in EvalFloat() local
|
| /external/tensorflow/tensorflow/lite/delegates/nnapi/ |
| D | nnapi_delegate_test.cc | 3092 LSTMOpModel(int n_batch, int n_input, int n_cell, int n_output, bool use_cifg, in LSTMOpModel() 3455 const int n_batch = 1; in TEST_F() local 3519 const int n_batch = 1; in TEST_F() local 3631 const int n_batch = 1; in TEST_F() local 4290 const int n_batch = 2; in TEST_F() local 4423 const int n_batch = 2; in TEST_F() local 4564 const int n_batch = 2; in TEST_F() local
|
| /external/deqp/external/openglcts/modules/gl/ |
| D | gl4cVertexAttrib64BitTest.cpp | 3554 for (unsigned int n_batch = 0; n_batch < m_n_batches; ++n_batch) in initBuffers() local 3883 for (unsigned int n_batch = 0; n_batch < m_n_batches; ++n_batch) in verifyXFBData() local
|