Searched refs:m_Params (Results 1 – 4 of 4) sorted by relevance
18 : m_Params(params) in ArmNNExecutor()28 auto network = parser->CreateNetwork(m_Params); in ArmNNExecutor()43 INetworkProperties networkProperties{m_Params.m_Concurrent, in ArmNNExecutor()64 if (m_Params.m_Iterations > 1) in ArmNNExecutor()67 msg << "Network will be executed " << m_Params.m_Iterations; in ArmNNExecutor()68 if (m_Params.m_Concurrent) in ArmNNExecutor()81 if (m_Params.m_GenerateTensorData) in ArmNNExecutor()86 if (m_Params.m_DontPrintOutputs) in ArmNNExecutor()100 for (size_t i = 0; i < m_Params.m_ThreadPoolSize; ++i) in ExecuteAsync()105 threadpool = std::make_unique<armnn::Threadpool>(m_Params.m_ThreadPoolSize, in ExecuteAsync()[all …]
10 : m_Params(params) in TfLiteExecutor()12 m_Model = tflite::FlatBufferModel::BuildFromFile(m_Params.m_ModelPath.c_str()); in TfLiteExecutor()15 LogAndThrow("Failed to load TfLite model from: " + m_Params.m_ModelPath); in TfLiteExecutor()29 if (m_Params.m_TfLiteExecutor == ExecuteNetworkParams::TfLiteExecutor::ArmNNTfLiteDelegate) in TfLiteExecutor()33 armnnDelegate::DelegateOptions delegateOptions = m_Params.ToDelegateOptions(); in TfLiteExecutor()53 armnn::Optional<std::string> dataFile = m_Params.m_GenerateTensorData in TfLiteExecutor()55 : armnn::MakeOptional<std::string>(m_Params.m_InputTensorDataFilePaths[inputIndex]); in TfLiteExecutor()122 for (size_t x = 0; x < m_Params.m_Iterations; x++) in Execute()130 if (!m_Params.m_DontPrintOutputs) in Execute()139 if (!m_Params.m_OutputTensorFiles.empty()) in Execute()[all …]
24 …TfLiteExecutor(const ExecuteNetworkParams& m_Params, armnn::IRuntime::CreationOptions runtimeOptio…32 const ExecuteNetworkParams& m_Params; member in TfLiteExecutor
116 ExecuteNetworkParams m_Params; member in ArmNNExecutor