From c5ee0d7460f1e0ec7e2b0639e3e8962934c4df09 Mon Sep 17 00:00:00 2001 From: John Mcloughlin Date: Fri, 24 Mar 2023 12:07:25 +0000 Subject: IVGCVSW-7197 Implement Pimpl Idiom for OptimizerOptions Signed-off-by: John Mcloughlin Change-Id: Id4bdc31e3e6f18ccaef232c29a2d2825c915b21c --- samples/CustomMemoryAllocatorSample.cpp | 6 +++--- .../ObjectDetection/include/delegate/ArmnnNetworkExecutor.hpp | 10 +++++----- 2 files changed, 8 insertions(+), 8 deletions(-) (limited to 'samples') diff --git a/samples/CustomMemoryAllocatorSample.cpp b/samples/CustomMemoryAllocatorSample.cpp index da249e0f4d..14c779e75f 100644 --- a/samples/CustomMemoryAllocatorSample.cpp +++ b/samples/CustomMemoryAllocatorSample.cpp @@ -1,5 +1,5 @@ // -// Copyright © 2021 Arm Ltd and Contributors. All rights reserved. +// Copyright © 2021, 2023 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // @@ -121,8 +121,8 @@ int main() fullyConnectedLayer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); // Optimise ArmNN network - OptimizerOptions optOptions; - optOptions.m_ImportEnabled = true; + OptimizerOptionsOpaque optOptions; + optOptions.SetImportEnabled(true); IOptimizedNetworkPtr optNet = Optimize(*network, {"GpuAcc"}, runtime->GetDeviceSpec(), optOptions); if (!optNet) diff --git a/samples/ObjectDetection/include/delegate/ArmnnNetworkExecutor.hpp b/samples/ObjectDetection/include/delegate/ArmnnNetworkExecutor.hpp index c8875a27dc..557ec8a8ac 100644 --- a/samples/ObjectDetection/include/delegate/ArmnnNetworkExecutor.hpp +++ b/samples/ObjectDetection/include/delegate/ArmnnNetworkExecutor.hpp @@ -1,5 +1,5 @@ // -// Copyright © 2022 Arm Ltd and Contributors. All rights reserved. +// Copyright © 2022, 2023 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // @@ -104,7 +104,7 @@ ArmnnNetworkExecutor::ArmnnNetworkExecutor(std::string& modelPath, m_profiling(isProfilingEnabled) { m_profiling.ProfilingStart(); - armnn::OptimizerOptions optimizerOptions; + armnn::OptimizerOptionsOpaque optimizerOptions; m_model = tflite::FlatBufferModel::BuildFromFile(modelPath.c_str()); if (m_model == nullptr) { @@ -130,12 +130,12 @@ ArmnnNetworkExecutor::ArmnnNetworkExecutor(std::string& modelPath, /* enable fast math optimization */ armnn::BackendOptions modelOptionGpu("GpuAcc", {{"FastMathEnabled", true}}); - optimizerOptions.m_ModelOptions.push_back(modelOptionGpu); + optimizerOptions.AddModelOption(modelOptionGpu); armnn::BackendOptions modelOptionCpu("CpuAcc", {{"FastMathEnabled", true}}); - optimizerOptions.m_ModelOptions.push_back(modelOptionCpu); + optimizerOptions.AddModelOption(modelOptionCpu); /* enable reduce float32 to float16 optimization */ - optimizerOptions.m_ReduceFp32ToFp16 = true; + optimizerOptions.SetReduceFp32ToFp16(true); armnnDelegate::DelegateOptions delegateOptions(preferredBackends, optimizerOptions); -- cgit v1.2.1