diff options
Diffstat (limited to 'tests/ExecuteNetwork')
-rw-r--r-- | tests/ExecuteNetwork/ArmNNExecutor.cpp | 26 | ||||
-rw-r--r-- | tests/ExecuteNetwork/ExecuteNetworkParams.cpp | 28 |
2 files changed, 27 insertions, 27 deletions
diff --git a/tests/ExecuteNetwork/ArmNNExecutor.cpp b/tests/ExecuteNetwork/ArmNNExecutor.cpp index 29ef4c5186..ac857a90df 100644 --- a/tests/ExecuteNetwork/ArmNNExecutor.cpp +++ b/tests/ExecuteNetwork/ArmNNExecutor.cpp @@ -518,15 +518,15 @@ armnn::IOptimizedNetworkPtr ArmNNExecutor::OptimizeNetwork(armnn::INetwork* netw { armnn::IOptimizedNetworkPtr optNet{nullptr, [](armnn::IOptimizedNetwork*){}}; - armnn::OptimizerOptions options; - options.m_ReduceFp32ToFp16 = m_Params.m_EnableFp16TurboMode; - options.m_Debug = m_Params.m_PrintIntermediate; - options.m_DebugToFile = m_Params.m_PrintIntermediateOutputsToFile; - options.m_shapeInferenceMethod = m_Params.m_InferOutputShape ? - armnn::ShapeInferenceMethod::InferAndValidate : - armnn::ShapeInferenceMethod::ValidateOnly; - options.m_ProfilingEnabled = m_Params.m_EnableProfiling; - options.m_AllowExpandedDims = m_Params.m_AllowExpandedDims; + armnn::OptimizerOptionsOpaque options; + options.SetReduceFp32ToFp16(m_Params.m_EnableFp16TurboMode); + options.SetDebugEnabled(m_Params.m_PrintIntermediate); + options.SetDebugToFileEnabled(m_Params.m_PrintIntermediateOutputsToFile); + options.SetShapeInferenceMethod(m_Params.m_InferOutputShape ? + armnn::ShapeInferenceMethod::InferAndValidate : + armnn::ShapeInferenceMethod::ValidateOnly); + options.SetProfilingEnabled(m_Params.m_EnableProfiling); + options.SetAllowExpandedDims(m_Params.m_AllowExpandedDims); armnn::BackendOptions gpuAcc("GpuAcc", { @@ -541,8 +541,8 @@ armnn::IOptimizedNetworkPtr ArmNNExecutor::OptimizeNetwork(armnn::INetwork* netw { "FastMathEnabled", m_Params.m_EnableFastMath }, { "NumberOfThreads", m_Params.m_NumberOfThreads } }); - options.m_ModelOptions.push_back(gpuAcc); - options.m_ModelOptions.push_back(cpuAcc); + options.AddModelOption(gpuAcc); + options.AddModelOption(cpuAcc); // The shapeInferenceMethod and allowExpandedDims values have to be added to the model options // because these are what are passed to the OptimizeSubgraphViews method and are used to create // the new optimized INetwork that method uses @@ -550,12 +550,12 @@ armnn::IOptimizedNetworkPtr ArmNNExecutor::OptimizeNetwork(armnn::INetwork* netw { { "AllowExpandedDims", m_Params.m_AllowExpandedDims } }); - options.m_ModelOptions.push_back(allowExDimOpt); + options.AddModelOption(allowExDimOpt); armnn::BackendOptions shapeInferOpt("ShapeInferenceMethod", { { "InferAndValidate", m_Params.m_InferOutputShape } }); - options.m_ModelOptions.push_back(shapeInferOpt); + options.AddModelOption(shapeInferOpt); const auto optimization_start_time = armnn::GetTimeNow(); optNet = armnn::Optimize(*network, m_Params.m_ComputeDevices, m_Runtime->GetDeviceSpec(), options); diff --git a/tests/ExecuteNetwork/ExecuteNetworkParams.cpp b/tests/ExecuteNetwork/ExecuteNetworkParams.cpp index fbfd1bc936..3628fa4976 100644 --- a/tests/ExecuteNetwork/ExecuteNetworkParams.cpp +++ b/tests/ExecuteNetwork/ExecuteNetworkParams.cpp @@ -1,5 +1,5 @@ // -// Copyright © 2022 Arm Ltd and Contributors. All rights reserved. +// Copyright © 2022-2023 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // @@ -137,33 +137,33 @@ armnnDelegate::DelegateOptions ExecuteNetworkParams::ToDelegateOptions() const } // Optimizer options next. - armnn::OptimizerOptions optimizerOptions; - optimizerOptions.m_ReduceFp32ToFp16 = m_EnableFp16TurboMode; - optimizerOptions.m_Debug = m_PrintIntermediate; - optimizerOptions.m_DebugToFile = m_PrintIntermediateOutputsToFile; - optimizerOptions.m_ProfilingEnabled = m_EnableProfiling; - optimizerOptions.m_shapeInferenceMethod = armnn::ShapeInferenceMethod::ValidateOnly; + armnn::OptimizerOptionsOpaque optimizerOptions; + optimizerOptions.SetReduceFp32ToFp16(m_EnableFp16TurboMode); + optimizerOptions.SetDebugEnabled(m_PrintIntermediate); + optimizerOptions.SetDebugToFileEnabled(m_PrintIntermediateOutputsToFile); + optimizerOptions.SetProfilingEnabled(m_EnableProfiling); + optimizerOptions.SetShapeInferenceMethod(armnn::ShapeInferenceMethod::ValidateOnly); if (m_InferOutputShape) { - optimizerOptions.m_shapeInferenceMethod = armnn::ShapeInferenceMethod::InferAndValidate; + optimizerOptions.SetShapeInferenceMethod(armnn::ShapeInferenceMethod::InferAndValidate); armnn::BackendOptions networkOption("ShapeInferenceMethod", { {"InferAndValidate", true} }); - optimizerOptions.m_ModelOptions.push_back(networkOption); + optimizerOptions.AddModelOption(networkOption); } { armnn::BackendOptions option("GpuAcc", {{"FastMathEnabled", m_EnableFastMath}}); - optimizerOptions.m_ModelOptions.push_back(option); + optimizerOptions.AddModelOption(option); } { armnn::BackendOptions option("GpuAcc", {{"CachedNetworkFilePath", m_CachedNetworkFilePath}}); - optimizerOptions.m_ModelOptions.push_back(option); + optimizerOptions.AddModelOption(option); } { armnn::BackendOptions option("GpuAcc", {{"MLGOTuningFilePath", m_MLGOTuningFilePath}}); - optimizerOptions.m_ModelOptions.push_back(option); + optimizerOptions.AddModelOption(option); } armnn::BackendOptions cpuAcc("CpuAcc", @@ -171,14 +171,14 @@ armnnDelegate::DelegateOptions ExecuteNetworkParams::ToDelegateOptions() const { "FastMathEnabled", m_EnableFastMath }, { "NumberOfThreads", m_NumberOfThreads } }); - optimizerOptions.m_ModelOptions.push_back(cpuAcc); + optimizerOptions.AddModelOption(cpuAcc); if (m_AllowExpandedDims) { armnn::BackendOptions networkOption("AllowExpandedDims", { {"AllowExpandedDims", true} }); - optimizerOptions.m_ModelOptions.push_back(networkOption); + optimizerOptions.AddModelOption(networkOption); } delegateOptions.SetOptimizerOptions(optimizerOptions); return delegateOptions; |