Searched defs:tensor (Results  1 – 9 of 9) sorted by relevance
| /packages/modules/NeuralNetworks/runtime/test/specs/V1_3/ | 
| D | bidirectional_sequence_rnn_1_3.mod.py | 20 def convert_to_time_major(tensor, tensor_shape):  argument30 def reverse_batch_major(tensor, tensor_shape):  argument
 33 def split_tensor_in_two(tensor, tensor_shape):  argument
 
 | 
| D | bidirectional_sequence_rnn_state_output.mod.py | 20 def convert_to_time_major(tensor, tensor_shape):  argument31 def reverse_batch_major(tensor, tensor_shape):  argument
 35 def split_tensor_in_two(tensor, tensor_shape):  argument
 
 | 
| D | unidirectional_sequence_rnn.mod.py | 42 def convert_to_time_major(tensor, num_batches, max_time, input_size):  argument
 | 
| /packages/modules/NeuralNetworks/runtime/test/specs/V1_2/ | 
| D | bidirectional_sequence_rnn.mod.py | 20 def convert_to_time_major(tensor, tensor_shape):  argument30 def reverse_batch_major(tensor, tensor_shape):  argument
 33 def split_tensor_in_two(tensor, tensor_shape):  argument
 
 | 
| D | unidirectional_sequence_rnn.mod.py | 39 def convert_to_time_major(tensor, num_batches, max_time, input_size):  argument
 | 
| /packages/modules/NeuralNetworks/runtime/operation_converters/ | 
| D | SubGraphContext.cpp | 54 int SubGraphContext::addTensorFlatbuffer(TensorFlatbuffer tensor, int32_t operandIdx) {  in addTensorFlatbuffer()205     TensorFlatbuffer tensor = tflite::CreateTensorDirect(  in createTensorFlatbufferFromOperand()  local
 
 | 
| /packages/modules/NeuralNetworks/common/cpu_operations/ | 
| D | QuantizedLSTMTest.cpp | 226     Result setInputTensor(Execution* execution, int tensor, const std::vector<T>& data) {  in setInputTensor()230     Result setOutputTensor(Execution* execution, int tensor, std::vector<T>* data) {  in setOutputTensor()
 
 | 
| D | QLSTM.cpp | 36 inline bool hasTensor(IOperationExecutionContext* context, const uint32_t tensor) {  in hasTensor()58     for (const int tensor : requiredTensorInputs) {  in prepare()  local
 
 | 
| D | UnidirectionalSequenceLSTM.cpp | 40 inline bool hasTensor(IOperationExecutionContext* context, const uint32_t tensor) {  in hasTensor()
 |