Searched refs:tensorIndex (Results 1 – 7 of 7) sorted by relevance
/external/armnn/samples/ObjectDetection/include/delegate/ |
D | ArmnnNetworkExecutor.hpp | 83 float GetOutputQuantizationScale(int tensorIndex); 85 int GetOutputQuantizationOffset(int tensorIndex); 240 float ArmnnNetworkExecutor<Tout>::GetOutputQuantizationScale(int tensorIndex) in GetOutputQuantizationScale() argument 242 assert(m_interpreter->outputs().size() > tensorIndex); in GetOutputQuantizationScale() 243 return m_interpreter->tensor(m_interpreter->outputs()[tensorIndex])->params.scale; in GetOutputQuantizationScale() 247 int ArmnnNetworkExecutor<Tout>::GetOutputQuantizationOffset(int tensorIndex) in GetOutputQuantizationOffset() argument 249 assert(m_interpreter->outputs().size() > tensorIndex); in GetOutputQuantizationOffset() 250 return m_interpreter->tensor(m_interpreter->outputs()[tensorIndex])->params.zero_point; in GetOutputQuantizationOffset()
|
/external/armnn/samples/common/include/ArmnnUtils/ |
D | ArmnnNetworkExecutor.hpp | 77 float GetOutputQuantizationScale(int tensorIndex); 79 int GetOutputQuantizationOffset(int tensorIndex); 216 float ArmnnNetworkExecutor<Tout>::GetOutputQuantizationScale(int tensorIndex) in GetOutputQuantizationScale() argument 218 assert(this->m_outputLayerNamesList.size() > tensorIndex); in GetOutputQuantizationScale() 219 return this->m_outputBindingInfo[tensorIndex].second.GetQuantizationScale(); in GetOutputQuantizationScale() 223 int ArmnnNetworkExecutor<Tout>::GetOutputQuantizationOffset(int tensorIndex) in GetOutputQuantizationOffset() argument 225 assert(this->m_outputLayerNamesList.size() > tensorIndex); in GetOutputQuantizationOffset() 226 return this->m_outputBindingInfo[tensorIndex].second.GetQuantizationOffset(); in GetOutputQuantizationOffset()
|
/external/tensorflow/tensorflow/lite/python/ |
D | util.py | 590 if tensor_maps[i].tensorIndex in map_old_to_new_tensors: 591 tensor_maps[i].tensorIndex = ( 592 map_old_to_new_tensors[tensor_maps[i].tensorIndex]) 738 if signature_def.inputs[i].tensorIndex == op.inputs[0]: 739 signature_def.inputs[i].tensorIndex = op.outputs[0] 861 if signature_def.outputs[i].tensorIndex == op.outputs[0]: 862 signature_def.outputs[i].tensorIndex = op.inputs[0] 922 if output.tensorIndex == op.outputs[0]: 923 output.tensorIndex = op.inputs[0] 937 if output.tensorIndex == dequant_op.outputs[0]: [all …]
|
/external/armnn/src/armnnTfLiteParser/ |
D | TfLiteParser.cpp | 188 size_t tensorIndex, in CheckTensor() argument 200 if (tensorIndex >= model->subgraphs[subgraphIndex]->tensors.size()) in CheckTensor() 207 tensorIndex, in CheckTensor() 620 armnn::LayerBindingId GenerateLayerBindingId(size_t subgraphIndex, size_t tensorIndex) in GenerateLayerBindingId() argument 624 return static_cast<armnn::LayerBindingId>((tensorIndex<<8)+subgraphIndex); in GenerateLayerBindingId() 1043 …for (size_t tensorIndex = 0; tensorIndex < m_SubgraphConnections[subgraphIndex].size(); ++tensorIn… in CreateNetworkFromModel() local 1045 if (m_SubgraphConnections[subgraphIndex][tensorIndex].outputSlot != nullptr) in CreateNetworkFromModel() 1048 … inputSlotIdx < m_SubgraphConnections[subgraphIndex][tensorIndex].inputSlots.size(); in CreateNetworkFromModel() 1051 m_SubgraphConnections[subgraphIndex][tensorIndex].outputSlot->Connect( in CreateNetworkFromModel() 1052 … *(m_SubgraphConnections[subgraphIndex][tensorIndex].inputSlots[inputSlotIdx])); in CreateNetworkFromModel() [all …]
|
D | TfLiteParser.hpp | 198 … void RegisterProducerOfTensor(size_t subgraphIndex, size_t tensorIndex, armnn::IOutputSlot* slot); 199 … void RegisterConsumerOfTensor(size_t subgraphIndex, size_t tensorIndex, armnn::IInputSlot* slot); 256 bool ShouldConstantTensorBeCreated(unsigned int tensorIndex);
|
/external/tensorflow/tensorflow/lite/java/src/main/java/org/tensorflow/lite/ |
D | TensorImpl.java | 39 static TensorImpl fromIndex(long nativeInterpreterHandle, int tensorIndex) { in fromIndex() argument 40 return new TensorImpl(create(nativeInterpreterHandle, tensorIndex, /*subgraphIndex=*/ 0)); in fromIndex() 489 private static native long create(long interpreterHandle, int tensorIndex, int subgraphIndex); in create() argument
|
/external/armnn/src/armnnTfLiteParser/test/ |
D | ParserFlatbuffersFixture.hpp | 96 … for (unsigned int tensorIndex = 0; tensorIndex < subgraph->tensors.size(); ++tensorIndex) in MakeModelDynamic() local 98 if (std::find(inputIds.begin(), inputIds.end(), tensorIndex) != inputIds.end()) in MakeModelDynamic()
|