/external/armnn/src/backends/backendsCommon/ |
D | IBackendInternal.cpp | 26 const ModelOptions& modelOptions) const in CreateWorkloadFactory() 28 if (!modelOptions.empty()) in CreateWorkloadFactory() 30 for (auto optionsGroup : modelOptions) in CreateWorkloadFactory() 44 const ModelOptions& modelOptions) const in CreateWorkloadFactory() 46 if (!modelOptions.empty()) in CreateWorkloadFactory() 48 for (auto optionsGroup : modelOptions) in CreateWorkloadFactory() 62 const ModelOptions& modelOptions, in CreateWorkloadFactory() argument 68 return CreateWorkloadFactory(tensorHandleFactoryRegistry, modelOptions); in CreateWorkloadFactory() 88 …l::ILayerSupportSharedPtr IBackendInternal::GetLayerSupport(const ModelOptions& modelOptions) const in GetLayerSupport() 90 if (!modelOptions.empty()) in GetLayerSupport() [all …]
|
/external/tflite-support/tensorflow_lite_support/ios/task/text/nlclassifier/Tests/ |
D | TFLNLClassifierTest.swift | 26 var modelOptions:TFLNLClassifierOptions!; variable 29 modelOptions = TFLNLClassifierOptions() in setUp() 30 modelOptions.inputTensorName = "input_text" in setUp() 31 modelOptions.outputScoreTensorName = "probability" in setUp() 37 options: modelOptions) in testClassifyPositiveResult() 50 options: modelOptions) in testClassifyNegativeResult()
|
D | TFLNLClassifierTest.m | 23 @property(nonatomic, nullable) TFLNLClassifierOptions *modelOptions; property 34 self.modelOptions = [[TFLNLClassifierOptions alloc] init]; 35 [self.modelOptions setInputTensorName:@"input_text"]; 36 [self.modelOptions setOutputScoreTensorName:@"probability"]; 41 options:self.modelOptions]; 54 options:self.modelOptions];
|
/external/armnn/tests/TfLiteMobileNetSsd-Armnn/ |
D | TfLiteMobileNetSsd-Armnn.cpp | 44 typename Model::CommandLineOptions modelOptions) in main() 46 if (!ValidateDirectory(modelOptions.m_ModelDir)) in main() 53 modelOptions.m_ModelDir + "ssd_mobilenet_v1.tflite"; in main() 63 … modelParams.m_ComputeDevices = modelOptions.GetComputeDevicesAsBackendIds(); in main() 64 … modelParams.m_VisualizePostOptimizationModel = modelOptions.m_VisualizePostOptimizationModel; in main() 65 … modelParams.m_EnableFp16TurboMode = modelOptions.m_EnableFp16TurboMode; in main()
|
/external/armnn/src/backends/neon/ |
D | NeonBackend.hpp | 44 const ModelOptions& modelOptions) const override; 47 const ModelOptions& modelOptions) const override; 53 …IBackendInternal::ILayerSupportSharedPtr GetLayerSupport(const ModelOptions& modelOptions) const o… 56 const ModelOptions& modelOptions) const override; 63 const ModelOptions& modelOptions) const override;
|
D | NeonBackendModelContext.cpp | 34 NeonBackendModelContext::NeonBackendModelContext(const ModelOptions& modelOptions) in NeonBackendModelContext() argument 37 if (!modelOptions.empty()) in NeonBackendModelContext() 39 ParseOptions(modelOptions, "CpuAcc", [&](std::string name, const BackendOptions::Var& value) in NeonBackendModelContext()
|
D | NeonBackend.cpp | 64 …t IBackendInternal::IMemoryManagerSharedPtr& memoryManager, const ModelOptions& modelOptions) const in CreateWorkloadFactory() 67 …ointerDowncast<NeonMemoryManager>(memoryManager), CreateBackendSpecificModelContext(modelOptions)); in CreateWorkloadFactory() 90 … TensorHandleFactoryRegistry& tensorHandleFactoryRegistry, const ModelOptions& modelOptions) const in CreateWorkloadFactory() 104 …ointerDowncast<NeonMemoryManager>(memoryManager), CreateBackendSpecificModelContext(modelOptions)); in CreateWorkloadFactory() 119 const ModelOptions& modelOptions) const in CreateBackendSpecificModelContext() 121 return IBackendSpecificModelContextPtr{new NeonBackendModelContext{modelOptions}}; in CreateBackendSpecificModelContext() 133 …ternal::ILayerSupportSharedPtr NeonBackend::GetLayerSupport(const ModelOptions& modelOptions) const in GetLayerSupport() 137 new NeonLayerSupport(CreateBackendSpecificModelContext(modelOptions)) in GetLayerSupport() 143 const ModelOptions& modelOptions) const in OptimizeSubgraphView() 145 OptimizationViews optimizationViews(modelOptions); in OptimizeSubgraphView()
|
D | NeonWorkloadFactory.cpp | 44 const ModelOptions& modelOptions) in IsLayerSupported() argument 46 …rn IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported, modelOptions); in IsLayerSupported() 63 auto modelOptions = dynamic_cast<NeonBackendModelContext*>(m_ModelContextPtr.get()); in SetNumberOfThreads() local 64 auto numberOfThreads = modelOptions->GetNumberOfThreads(); in SetNumberOfThreads() 218 … auto modelOptions = dynamic_cast<NeonBackendModelContext*>(m_ModelContextPtr.get()); in CreateWorkload() local 219 if (modelOptions) in CreateWorkload() 221 isFastMathEnabled = modelOptions->IsFastMathEnabled(); in CreateWorkload() 239 … auto modelOptions = dynamic_cast<NeonBackendModelContext*>(m_ModelContextPtr.get()); in CreateWorkload() local 240 if (modelOptions) in CreateWorkload() 242 isFastMathEnabled = modelOptions->IsFastMathEnabled(); in CreateWorkload() [all …]
|
/external/armnn/include/armnn/backends/ |
D | IBackendInternal.hpp | 111 const ModelOptions& modelOptions) const; 115 const ModelOptions& modelOptions) const; 119 const ModelOptions& modelOptions, 132 …dSpecificModelContextPtr CreateBackendSpecificModelContext(const ModelOptions& modelOptions) const; 140 virtual ILayerSupportSharedPtr GetLayerSupport(const ModelOptions& modelOptions) const; 145 const ModelOptions& modelOptions) const;
|
/external/armnn/src/backends/cl/test/ |
D | ClBackendTests.cpp | 60 ModelOptions modelOptions; variable 61 clBackend->CreateWorkloadFactory(registry, modelOptions); 73 ModelOptions modelOptions; variable 74 clBackend->CreateWorkloadFactory(registry, modelOptions,
|
D | ClWorkloadFactoryHelper.hpp | 31 const armnn::ModelOptions& modelOptions = {}) in GetFactory() 35 backend.CreateBackendSpecificModelContext(modelOptions));
|
D | ClMemCopyTests.cpp | 27 const armnn::ModelOptions& modelOptions = {}) in GetFactory() 31 backend.CreateBackendSpecificModelContext(modelOptions));
|
/external/armnn/src/backends/cl/ |
D | ClBackendModelContext.cpp | 34 ClBackendModelContext::ClBackendModelContext(const ModelOptions& modelOptions) in ClBackendModelContext() argument 37 if (!modelOptions.empty()) in ClBackendModelContext() 39 ParseOptions(modelOptions, "GpuAcc", [&](std::string name, const BackendOptions::Var& value) in ClBackendModelContext()
|
D | ClWorkloadFactory.cpp | 56 const ModelOptions& modelOptions) in IsLayerSupported() argument 58 …rn IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported, modelOptions); in IsLayerSupported() 70 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get()); in AfterWorkloadsCreated() local 71 if (modelOptions->SaveCachedNetwork()) in AfterWorkloadsCreated() 75 auto cachedFd = modelOptions->GetCachedFileDescriptor(); in AfterWorkloadsCreated() 96 auto filePath = modelOptions->GetCachedNetworkFilePath(); in AfterWorkloadsCreated() 147 auto modelOptions = dynamic_cast<ClBackendModelContext*>(m_ModelContextPtr.get()); in InitializeCLCompileContext() local 148 auto filePath = modelOptions->GetCachedNetworkFilePath(); in InitializeCLCompileContext() 149 if (!(modelOptions->SaveCachedNetwork())) in InitializeCLCompileContext() 152 auto cachedFd = modelOptions->GetCachedFileDescriptor(); in InitializeCLCompileContext() [all …]
|
D | ClBackend.cpp | 68 …t IBackendInternal::IMemoryManagerSharedPtr& memoryManager, const ModelOptions& modelOptions) const in CreateWorkloadFactory() 71 …cPointerDowncast<ClMemoryManager>(memoryManager), CreateBackendSpecificModelContext(modelOptions)); in CreateWorkloadFactory() 103 TensorHandleFactoryRegistry& registry, const ModelOptions& modelOptions) const in CreateWorkloadFactory() 127 …cPointerDowncast<ClMemoryManager>(memoryManager), CreateBackendSpecificModelContext(modelOptions)); in CreateWorkloadFactory() 132 const ModelOptions& modelOptions, in CreateWorkloadFactory() argument 167 …cPointerDowncast<ClMemoryManager>(memoryManager), CreateBackendSpecificModelContext(modelOptions)); in CreateWorkloadFactory() 248 const ModelOptions& modelOptions) const in CreateBackendSpecificModelContext() 250 return IBackendSpecificModelContextPtr{new ClBackendModelContext{modelOptions}}; in CreateBackendSpecificModelContext() 262 …Internal::ILayerSupportSharedPtr ClBackend::GetLayerSupport(const ModelOptions& modelOptions) const in GetLayerSupport() 266 new ClLayerSupport(CreateBackendSpecificModelContext(modelOptions)) in GetLayerSupport() [all …]
|
D | ClBackend.hpp | 60 const ModelOptions& modelOptions) const override; 63 const ModelOptions& modelOptions) const override; 66 const ModelOptions& modelOptions, 83 …IBackendInternal::ILayerSupportSharedPtr GetLayerSupport(const ModelOptions& modelOptions) const o… 86 const ModelOptions& modelOptions) const override; 89 const ModelOptions& modelOptions) const override;
|
/external/armnn/src/backends/aclCommon/test/ |
D | MemCopyTests.cpp | 59 const armnn::ModelOptions& modelOptions = {}) in GetFactory() 63 backend.CreateBackendSpecificModelContext(modelOptions)); 78 const armnn::ModelOptions& modelOptions = {}) in GetFactory() 82 backend.CreateBackendSpecificModelContext(modelOptions));
|
/external/armnn/src/backends/neon/test/ |
D | NeonOptimizedNetworkTests.cpp | 100 armnn::BackendOptions modelOptions("CpuAcc", {{"FastMathEnabled", true}}); variable 101 optimizerOptions.AddModelOption(modelOptions); 132 armnn::BackendOptions modelOptions("CpuAcc", {{"NumberOfThreads", numberOfThreads}}); variable 133 optimizerOptions.AddModelOption(modelOptions);
|
D | NeonWorkloadFactoryHelper.hpp | 31 const armnn::ModelOptions& modelOptions = {}) in GetFactory() 35 backend.CreateBackendSpecificModelContext(modelOptions));
|
D | NeonBackendTests.cpp | 44 ModelOptions modelOptions; variable 45 neonBackend->CreateWorkloadFactory(registry, modelOptions);
|
D | NeonMemCopyTests.cpp | 27 const armnn::ModelOptions& modelOptions = {}) in GetFactory() 31 backend.CreateBackendSpecificModelContext(modelOptions));
|
/external/armnn/include/armnn/ |
D | INetwork.hpp | 157 … ModelOptions modelOptions = {}, bool exportEnabled = false, bool debugToFile = false) in OptimizerOptions() 164 , m_ModelOptions(modelOptions) 174 … bool importEnabled = false, ModelOptions modelOptions = {}, bool exportEnabled = false, in OptimizerOptions() 182 , m_ModelOptions(modelOptions) 272 … ModelOptions modelOptions = {}, bool exportEnabled = false, bool debugToFile = false); 276 … bool importEnabled = false, ModelOptions modelOptions = {}, bool exportEnabled = false, 879 IOptimizedNetwork(const IOptimizedNetwork& other, const ModelOptions& modelOptions); 905 IOptimizedNetwork(std::unique_ptr<Graph> graph, const ModelOptions& modelOptions);
|
/external/armnn/src/armnn/ |
D | OptimizedNetworkImpl.hpp | 14 OptimizedNetworkImpl(const OptimizedNetworkImpl& other, const ModelOptions& modelOptions); 16 OptimizedNetworkImpl(std::unique_ptr<Graph> graph, const ModelOptions& modelOptions);
|
D | Network.hpp | 315 bool importEnabled, ModelOptions modelOptions = {}, in OptimizerOptionsOpaqueImpl() 323 , m_ModelOptions(modelOptions) 332 … bool importEnabled, ModelOptions modelOptions, bool exportEnabled, in OptimizerOptionsOpaqueImpl() 340 , m_ModelOptions(modelOptions) in OptimizerOptionsOpaqueImpl()
|
/external/armnn/tests/ |
D | InferenceTest.inl | 387 typename InferenceModel::CommandLineOptions modelOptions) 389 if (!ValidateDirectory(modelOptions.m_ModelDir)) 395 modelParams.m_ModelPath = modelOptions.m_ModelDir + modelFilename; 405 modelParams.m_ComputeDevices = modelOptions.GetComputeDevicesAsBackendIds(); 406 … modelParams.m_VisualizePostOptimizationModel = modelOptions.m_VisualizePostOptimizationModel; 407 modelParams.m_EnableFp16TurboMode = modelOptions.m_EnableFp16TurboMode;
|