Home
last modified time | relevance | path

Searched refs:tensorIndex (Results 1 – 7 of 7) sorted by relevance

/external/armnn/samples/ObjectDetection/include/delegate/
DArmnnNetworkExecutor.hpp83 float GetOutputQuantizationScale(int tensorIndex);
85 int GetOutputQuantizationOffset(int tensorIndex);
240 float ArmnnNetworkExecutor<Tout>::GetOutputQuantizationScale(int tensorIndex) in GetOutputQuantizationScale() argument
242 assert(m_interpreter->outputs().size() > tensorIndex); in GetOutputQuantizationScale()
243 return m_interpreter->tensor(m_interpreter->outputs()[tensorIndex])->params.scale; in GetOutputQuantizationScale()
247 int ArmnnNetworkExecutor<Tout>::GetOutputQuantizationOffset(int tensorIndex) in GetOutputQuantizationOffset() argument
249 assert(m_interpreter->outputs().size() > tensorIndex); in GetOutputQuantizationOffset()
250 return m_interpreter->tensor(m_interpreter->outputs()[tensorIndex])->params.zero_point; in GetOutputQuantizationOffset()
/external/armnn/samples/common/include/ArmnnUtils/
DArmnnNetworkExecutor.hpp77 float GetOutputQuantizationScale(int tensorIndex);
79 int GetOutputQuantizationOffset(int tensorIndex);
216 float ArmnnNetworkExecutor<Tout>::GetOutputQuantizationScale(int tensorIndex) in GetOutputQuantizationScale() argument
218 assert(this->m_outputLayerNamesList.size() > tensorIndex); in GetOutputQuantizationScale()
219 return this->m_outputBindingInfo[tensorIndex].second.GetQuantizationScale(); in GetOutputQuantizationScale()
223 int ArmnnNetworkExecutor<Tout>::GetOutputQuantizationOffset(int tensorIndex) in GetOutputQuantizationOffset() argument
225 assert(this->m_outputLayerNamesList.size() > tensorIndex); in GetOutputQuantizationOffset()
226 return this->m_outputBindingInfo[tensorIndex].second.GetQuantizationOffset(); in GetOutputQuantizationOffset()
/external/tensorflow/tensorflow/lite/python/
Dutil.py590 if tensor_maps[i].tensorIndex in map_old_to_new_tensors:
591 tensor_maps[i].tensorIndex = (
592 map_old_to_new_tensors[tensor_maps[i].tensorIndex])
738 if signature_def.inputs[i].tensorIndex == op.inputs[0]:
739 signature_def.inputs[i].tensorIndex = op.outputs[0]
861 if signature_def.outputs[i].tensorIndex == op.outputs[0]:
862 signature_def.outputs[i].tensorIndex = op.inputs[0]
922 if output.tensorIndex == op.outputs[0]:
923 output.tensorIndex = op.inputs[0]
937 if output.tensorIndex == dequant_op.outputs[0]:
[all …]
/external/armnn/src/armnnTfLiteParser/
DTfLiteParser.cpp188 size_t tensorIndex, in CheckTensor() argument
200 if (tensorIndex >= model->subgraphs[subgraphIndex]->tensors.size()) in CheckTensor()
207 tensorIndex, in CheckTensor()
620 armnn::LayerBindingId GenerateLayerBindingId(size_t subgraphIndex, size_t tensorIndex) in GenerateLayerBindingId() argument
624 return static_cast<armnn::LayerBindingId>((tensorIndex<<8)+subgraphIndex); in GenerateLayerBindingId()
1043 …for (size_t tensorIndex = 0; tensorIndex < m_SubgraphConnections[subgraphIndex].size(); ++tensorIn… in CreateNetworkFromModel() local
1045 if (m_SubgraphConnections[subgraphIndex][tensorIndex].outputSlot != nullptr) in CreateNetworkFromModel()
1048 … inputSlotIdx < m_SubgraphConnections[subgraphIndex][tensorIndex].inputSlots.size(); in CreateNetworkFromModel()
1051 m_SubgraphConnections[subgraphIndex][tensorIndex].outputSlot->Connect( in CreateNetworkFromModel()
1052 … *(m_SubgraphConnections[subgraphIndex][tensorIndex].inputSlots[inputSlotIdx])); in CreateNetworkFromModel()
[all …]
DTfLiteParser.hpp198 … void RegisterProducerOfTensor(size_t subgraphIndex, size_t tensorIndex, armnn::IOutputSlot* slot);
199 … void RegisterConsumerOfTensor(size_t subgraphIndex, size_t tensorIndex, armnn::IInputSlot* slot);
256 bool ShouldConstantTensorBeCreated(unsigned int tensorIndex);
/external/tensorflow/tensorflow/lite/java/src/main/java/org/tensorflow/lite/
DTensorImpl.java39 static TensorImpl fromIndex(long nativeInterpreterHandle, int tensorIndex) { in fromIndex() argument
40 return new TensorImpl(create(nativeInterpreterHandle, tensorIndex, /*subgraphIndex=*/ 0)); in fromIndex()
489 private static native long create(long interpreterHandle, int tensorIndex, int subgraphIndex); in create() argument
/external/armnn/src/armnnTfLiteParser/test/
DParserFlatbuffersFixture.hpp96 … for (unsigned int tensorIndex = 0; tensorIndex < subgraph->tensors.size(); ++tensorIndex) in MakeModelDynamic() local
98 if (std::find(inputIds.begin(), inputIds.end(), tensorIndex) != inputIds.end()) in MakeModelDynamic()