aboutsummaryrefslogtreecommitdiff
path: root/tests/ExecuteNetwork/ExecuteNetworkParams.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'tests/ExecuteNetwork/ExecuteNetworkParams.cpp')
-rw-r--r--tests/ExecuteNetwork/ExecuteNetworkParams.cpp134
1 files changed, 3 insertions, 131 deletions
diff --git a/tests/ExecuteNetwork/ExecuteNetworkParams.cpp b/tests/ExecuteNetwork/ExecuteNetworkParams.cpp
index cc75bb4323..f341c30738 100644
--- a/tests/ExecuteNetwork/ExecuteNetworkParams.cpp
+++ b/tests/ExecuteNetwork/ExecuteNetworkParams.cpp
@@ -1,76 +1,15 @@
//
-// Copyright © 2020 Arm Ltd and Contributors. All rights reserved.
+// Copyright © 2022 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
#include "ExecuteNetworkParams.hpp"
#include "NetworkExecutionUtils/NetworkExecutionUtils.hpp"
-#include <InferenceModel.hpp>
#include <armnn/Logging.hpp>
#include <fmt/format.h>
-
-bool IsModelBinary(const std::string& modelFormat)
-{
- // Parse model binary flag from the model-format string we got from the command-line
- if (modelFormat.find("binary") != std::string::npos)
- {
- return true;
- }
- else if (modelFormat.find("txt") != std::string::npos || modelFormat.find("text") != std::string::npos)
- {
- return false;
- }
- else
- {
- throw armnn::InvalidArgumentException(fmt::format("Unknown model format: '{}'. "
- "Please include 'binary' or 'text'",
- modelFormat));
- }
-}
-
-void CheckModelFormat(const std::string& modelFormat)
-{
- // Forward to implementation based on the parser type
- if (modelFormat.find("armnn") != std::string::npos)
- {
-#if defined(ARMNN_SERIALIZER)
-#else
- throw armnn::InvalidArgumentException("Can't run model in armnn format without a "
- "built with serialization support.");
-#endif
- }
- else if (modelFormat.find("onnx") != std::string::npos)
- {
-#if defined(ARMNN_ONNX_PARSER)
-#else
- throw armnn::InvalidArgumentException("Can't run model in onnx format without a "
- "built with Onnx parser support.");
-#endif
- }
- else if (modelFormat.find("tflite") != std::string::npos)
- {
-#if defined(ARMNN_TF_LITE_PARSER)
- if (!IsModelBinary(modelFormat))
- {
- throw armnn::InvalidArgumentException(fmt::format("Unknown model format: '{}'. Only 'binary' "
- "format supported for tflite files",
- modelFormat));
- }
-#elif defined(ARMNN_TFLITE_DELEGATE)
-#else
- throw armnn::InvalidArgumentException("Can't run model in tflite format without a "
- "built with Tensorflow Lite parser support.");
-#endif
- }
- else
- {
- throw armnn::InvalidArgumentException(fmt::format("Unknown model format: '{}'. "
- "Please include 'tflite' or 'onnx'",
- modelFormat));
- }
-}
+#include <armnnUtils/Filesystem.hpp>
void CheckClTuningParameter(const int& tuningLevel,
const std::string& tuningPath,
@@ -105,7 +44,6 @@ void CheckClTuningParameter(const int& tuningLevel,
ARMNN_LOG(warning) << "To use Cl Tuning the compute device GpuAcc needs to be active.";
}
}
-
}
void ExecuteNetworkParams::ValidateParams()
@@ -120,7 +58,6 @@ void ExecuteNetworkParams::ValidateParams()
<< invalidBackends;
}
}
-
CheckClTuningParameter(m_TuningLevel, m_TuningPath, m_ComputeDevices);
if (m_EnableBf16TurboMode && m_EnableFp16TurboMode)
@@ -129,10 +66,6 @@ void ExecuteNetworkParams::ValidateParams()
"enabled at the same time.");
}
- m_IsModelBinary = IsModelBinary(m_ModelFormat);
-
- CheckModelFormat(m_ModelFormat);
-
// Check input tensor shapes
if ((m_InputTensorShapes.size() != 0) &&
(m_InputTensorShapes.size() != m_InputNames.size()))
@@ -157,68 +90,6 @@ void ExecuteNetworkParams::ValidateParams()
m_InputNames.size(),
m_InputTensorDataFilePaths.size()));
}
- else if (m_InputTensorDataFilePaths.size() % m_InputNames.size() != 0)
- {
- throw armnn::InvalidArgumentException(
- fmt::format("According to the number of input names the user provided the network has {} "
- "inputs. The user specified {} input-tensor-data file paths which is not "
- "divisible by the number of inputs.",
- m_InputNames.size(),
- m_InputTensorDataFilePaths.size()));
- }
- }
-
- if (m_InputTypes.size() == 0)
- {
- //Defaults the value of all inputs to "float"
- m_InputTypes.assign(m_InputNames.size(), "float");
- }
- else if ((m_InputTypes.size() != 0) &&
- (m_InputTypes.size() != m_InputNames.size()))
- {
- throw armnn::InvalidArgumentException("input-name and input-type must have the same amount of elements.");
- }
-
- // Make sure that the number of input files given is divisible by the number of inputs of the model
- if (!(m_InputTensorDataFilePaths.size() % m_InputNames.size() == 0))
- {
- throw armnn::InvalidArgumentException(
- fmt::format("The number of input-tensor-data files ({0}) is not divisible by the "
- "number of inputs ({1} according to the number of input names).",
- m_InputTensorDataFilePaths.size(),
- m_InputNames.size()));
- }
-
- if (m_OutputTypes.size() == 0)
- {
- //Defaults the value of all outputs to "float"
- m_OutputTypes.assign(m_OutputNames.size(), "float");
- }
- else if ((m_OutputTypes.size() != 0) &&
- (m_OutputTypes.size() != m_OutputNames.size()))
- {
- throw armnn::InvalidArgumentException("output-name and output-type must have the same amount of elements.");
- }
-
- // Make sure that the number of output files given is equal to the number of outputs of the model
- // or equal to the number of outputs of the model multiplied with the number of iterations
- if (!m_OutputTensorFiles.empty())
- {
- if ((m_OutputTensorFiles.size() != m_OutputNames.size()) &&
- (m_OutputTensorFiles.size() != m_OutputNames.size() * m_Iterations))
- {
- std::stringstream errmsg;
- auto numOutputs = m_OutputNames.size();
- throw armnn::InvalidArgumentException(
- fmt::format("The user provided {0} output-tensor files. The only allowed number of output-tensor "
- "files is the number of outputs of the network ({1} according to the number of "
- "output names) or the number of outputs multiplied with the number of times the "
- "network should be executed (NumOutputs * NumIterations = {1} * {2} = {3}).",
- m_OutputTensorFiles.size(),
- numOutputs,
- m_Iterations,
- numOutputs*m_Iterations));
- }
}
// Check that threshold time is not less than zero
@@ -310,4 +181,5 @@ armnnDelegate::DelegateOptions ExecuteNetworkParams::ToDelegateOptions() const
return delegateOptions;
}
+
#endif