Lines Matching refs:m_Params
10 : m_Params(params) in TfLiteExecutor()
12 m_Model = tflite::FlatBufferModel::BuildFromFile(m_Params.m_ModelPath.c_str()); in TfLiteExecutor()
15 LogAndThrow("Failed to load TfLite model from: " + m_Params.m_ModelPath); in TfLiteExecutor()
29 if (m_Params.m_TfLiteExecutor == ExecuteNetworkParams::TfLiteExecutor::ArmNNTfLiteDelegate) in TfLiteExecutor()
33 armnnDelegate::DelegateOptions delegateOptions = m_Params.ToDelegateOptions(); in TfLiteExecutor()
53 armnn::Optional<std::string> dataFile = m_Params.m_GenerateTensorData in TfLiteExecutor()
55 : armnn::MakeOptional<std::string>(m_Params.m_InputTensorDataFilePaths[inputIndex]); in TfLiteExecutor()
122 for (size_t x = 0; x < m_Params.m_Iterations; x++) in Execute()
130 if (!m_Params.m_DontPrintOutputs) in Execute()
139 if (!m_Params.m_OutputTensorFiles.empty()) in Execute()
141 … outputTensorFile = fopen(m_Params.m_OutputTensorFiles[outputIndex].c_str(), "w"); in Execute()
144 … LogAndThrow("Specified output tensor file, \"" + m_Params.m_OutputTensorFiles[outputIndex] + in Execute()
150 … << " to file: '" << m_Params.m_OutputTensorFiles[outputIndex] << "'"; in Execute()
224 CheckInferenceTimeThreshold(duration, m_Params.m_ThresholdTime); in Execute()