diff options
author | Narumol Prangnawarat <narumol.prangnawarat@arm.com> | 2020-03-24 13:54:05 +0000 |
---|---|---|
committer | Jim Flynn <jim.flynn@arm.com> | 2020-03-24 16:45:36 +0000 |
commit | d8cc8116f2deea11ad7aff9218a2e103062a7daf (patch) | |
tree | 63ca615ca9d5f8a1655f560518fc65b662d65e51 /tests/NetworkExecutionUtils | |
parent | b6a402f46231688f7684dcb8c8e4ef5f4579b011 (diff) | |
download | armnn-d8cc8116f2deea11ad7aff9218a2e103062a7daf.tar.gz |
IVGCVSW-4521 Add bf16-turbo-mode option to ExecuteNetwork
Signed-off-by: Narumol Prangnawarat <narumol.prangnawarat@arm.com>
Change-Id: I57ec47adf98680254fa481fb91d5a98dea8f032e
Diffstat (limited to 'tests/NetworkExecutionUtils')
-rw-r--r-- | tests/NetworkExecutionUtils/NetworkExecutionUtils.hpp | 13 |
1 files changed, 9 insertions, 4 deletions
diff --git a/tests/NetworkExecutionUtils/NetworkExecutionUtils.hpp b/tests/NetworkExecutionUtils/NetworkExecutionUtils.hpp index 4d996fd401..a0aeb8bc5a 100644 --- a/tests/NetworkExecutionUtils/NetworkExecutionUtils.hpp +++ b/tests/NetworkExecutionUtils/NetworkExecutionUtils.hpp @@ -379,6 +379,7 @@ struct ExecuteNetworkParams bool m_DequantizeOutput; bool m_EnableProfiling; bool m_EnableFp16TurboMode; + bool m_EnableBf16TurboMode; double m_ThresholdTime; bool m_PrintIntermediate; size_t m_SubgraphId; @@ -424,6 +425,7 @@ int MainImpl(const ExecuteNetworkParams& params, inferenceModelParams.m_SubgraphId = params.m_SubgraphId; inferenceModelParams.m_EnableFp16TurboMode = params.m_EnableFp16TurboMode; + inferenceModelParams.m_EnableBf16TurboMode = params.m_EnableBf16TurboMode; InferenceModel<TParser, TDataType> model(inferenceModelParams, params.m_EnableProfiling, @@ -549,6 +551,7 @@ int RunTest(const std::string& format, bool dequantizeOuput, bool enableProfiling, bool enableFp16TurboMode, + bool enableBf16TurboMode, const double& thresholdTime, bool printIntermediate, const size_t subgraphId, @@ -673,6 +676,7 @@ int RunTest(const std::string& format, params.m_DequantizeOutput = dequantizeOuput; params.m_EnableProfiling = enableProfiling; params.m_EnableFp16TurboMode = enableFp16TurboMode; + params.m_EnableBf16TurboMode = enableBf16TurboMode; params.m_ThresholdTime = thresholdTime; params.m_PrintIntermediate = printIntermediate; params.m_SubgraphId = subgraphId; @@ -748,8 +752,9 @@ int RunTest(const std::string& format, } int RunCsvTest(const armnnUtils::CsvRow &csvRow, const std::shared_ptr<armnn::IRuntime>& runtime, - const bool enableProfiling, const bool enableFp16TurboMode, const double& thresholdTime, - const bool printIntermediate, bool enableLayerDetails = false, bool parseUnuspported = false) + const bool enableProfiling, const bool enableFp16TurboMode, const bool enableBf16TurboMode, + const double& thresholdTime, const bool printIntermediate, bool enableLayerDetails = false, + bool parseUnuspported = false) { IgnoreUnused(runtime); std::string modelFormat; @@ -868,6 +873,6 @@ int RunCsvTest(const armnnUtils::CsvRow &csvRow, const std::shared_ptr<armnn::IR return RunTest(modelFormat, inputTensorShapes, computeDevices, dynamicBackendsPath, modelPath, inputNames, inputTensorDataFilePaths, inputTypes, quantizeInput, outputTypes, outputNames, outputTensorFiles, - dequantizeOutput, enableProfiling, enableFp16TurboMode, thresholdTime, printIntermediate, subgraphId, - enableLayerDetails, parseUnuspported); + dequantizeOutput, enableProfiling, enableFp16TurboMode, enableBf16TurboMode, + thresholdTime, printIntermediate, subgraphId, enableLayerDetails, parseUnuspported); } |