From c5ee0d7460f1e0ec7e2b0639e3e8962934c4df09 Mon Sep 17 00:00:00 2001 From: John Mcloughlin Date: Fri, 24 Mar 2023 12:07:25 +0000 Subject: IVGCVSW-7197 Implement Pimpl Idiom for OptimizerOptions Signed-off-by: John Mcloughlin Change-Id: Id4bdc31e3e6f18ccaef232c29a2d2825c915b21c --- delegate/test/DelegateOptionsTest.cpp | 17 ++++++++++------- 1 file changed, 10 insertions(+), 7 deletions(-) (limited to 'delegate/test/DelegateOptionsTest.cpp') diff --git a/delegate/test/DelegateOptionsTest.cpp b/delegate/test/DelegateOptionsTest.cpp index d84d420977..fd1ef88645 100644 --- a/delegate/test/DelegateOptionsTest.cpp +++ b/delegate/test/DelegateOptionsTest.cpp @@ -26,7 +26,7 @@ TEST_CASE ("ArmnnDelegateOptimizerOptionsReduceFp32ToFp16") std::vector expectedResult = { 1, 2, 2, 2 }; // Enable ReduceFp32ToFp16 - armnn::OptimizerOptions optimizerOptions(true, true, false, false); + armnn::OptimizerOptionsOpaque optimizerOptions(true, true, false, false); armnnDelegate::DelegateOptions delegateOptions(backends, optimizerOptions); DelegateOptionTest(::tflite::TensorType_FLOAT32, @@ -55,7 +55,7 @@ TEST_CASE ("ArmnnDelegateOptimizerOptionsDebug") std::vector expectedResult = { 1, 2, 2, 2 }; // Enable Debug - armnn::OptimizerOptions optimizerOptions(false, true, false, false); + armnn::OptimizerOptionsOpaque optimizerOptions(false, true, false, false); armnnDelegate::DelegateOptions delegateOptions(backends, optimizerOptions); DelegateOptionTest(::tflite::TensorType_FLOAT32, @@ -83,7 +83,7 @@ TEST_CASE ("ArmnnDelegateOptimizerOptionsDebugFunction") std::vector expectedResult = { 1, 2, 2, 2 }; // Enable debug with debug callback function - armnn::OptimizerOptions optimizerOptions(false, true, false, false); + armnn::OptimizerOptionsOpaque optimizerOptions(false, true, false, false); bool callback = false; auto mockCallback = [&](LayerGuid guid, unsigned int slotIndex, armnn::ITensorHandle* tensor) { @@ -121,7 +121,7 @@ TEST_CASE ("ArmnnDelegateOptimizerOptionsImport") std::vector divData = { 2, 2, 3, 4 }; std::vector expectedResult = { 1, 2, 2, 2 }; - armnn::OptimizerOptions optimizerOptions(false, false, false, true); + armnn::OptimizerOptionsOpaque optimizerOptions(false, false, false, true); armnnDelegate::DelegateOptions delegateOptions(backends, optimizerOptions); DelegateOptionTest(::tflite::TensorType_UINT8, @@ -227,7 +227,8 @@ TEST_CASE ("ArmnnDelegateModelOptions_CpuAcc_Test") }); modelOptions.push_back(cpuAcc); - armnn::OptimizerOptions optimizerOptions(false, false, false, false, modelOptions, false); + armnn::OptimizerOptionsOpaque optimizerOptions(false, false, false, + false, modelOptions, false); armnnDelegate::DelegateOptions delegateOptions(backends, optimizerOptions); DelegateOptionTest(::tflite::TensorType_FLOAT32, @@ -256,7 +257,8 @@ TEST_CASE ("ArmnnDelegateSerializeToDot") std::vector divData = { 2, 2, 3, 4 }; std::vector expectedResult = { 1, 2, 2, 2 }; - armnn::OptimizerOptions optimizerOptions(false, false, false, false); + armnn::OptimizerOptionsOpaque optimizerOptions(false, false, + false, false); armnnDelegate::DelegateOptions delegateOptions(backends, optimizerOptions); // Enable serialize to dot by specifying the target file name. delegateOptions.SetSerializeToDot(filename); @@ -299,7 +301,8 @@ void CreateFp16StringParsingTestRun(std::vector& keys, options_values.get()[i] = values[i].c_str(); } - armnnDelegate::DelegateOptions delegateOptions(options_keys.get(), options_values.get(), num_options, nullptr); + armnnDelegate::DelegateOptions delegateOptions(options_keys.get(), options_values.get(), + num_options, nullptr); DelegateOptionTest(::tflite::TensorType_FLOAT32, tensorShape, inputData, -- cgit v1.2.1