From 8800c00770ed14eb48045cfcf033d6b67595a126 Mon Sep 17 00:00:00 2001 From: Matthew Bentham Date: Mon, 19 Nov 2018 13:19:28 +0000 Subject: IVGCVSW-2169 Remove DataLayoutIndexed from public API Change-Id: If8d8087d9d365e467d3ca9bf9c40d7219cb75cfd --- .../backendsCommon/test/Conv2dTestImpl.hpp | 41 +++++++++++----------- 1 file changed, 21 insertions(+), 20 deletions(-) (limited to 'src/backends/backendsCommon/test/Conv2dTestImpl.hpp') diff --git a/src/backends/backendsCommon/test/Conv2dTestImpl.hpp b/src/backends/backendsCommon/test/Conv2dTestImpl.hpp index d99b7f7fa3..6685a8edd2 100755 --- a/src/backends/backendsCommon/test/Conv2dTestImpl.hpp +++ b/src/backends/backendsCommon/test/Conv2dTestImpl.hpp @@ -14,6 +14,7 @@ #include #include "QuantizeHelper.hpp" +#include #include #include #include @@ -75,7 +76,7 @@ LayerTestResult SimpleConvolution2dTestImpl( const boost::multi_array& originalOutputExpected, float qScale, int32_t qOffset, - const armnn::DataLayoutIndexed& layout = armnn::DataLayout::NCHW, + const armnn::DataLayout layout = armnn::DataLayout::NCHW, uint32_t padLeft = 0, uint32_t padTop = 0, uint32_t padRight = 0, @@ -137,7 +138,7 @@ LayerTestResult SimpleConvolution2dTestImpl( // at this point if we require it permute the input data const armnn::PermutationVector NCHWToNHWC = { 0, 3, 1, 2 }; - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { std::vector tmp(inputData.size()); armnnUtils::Permute(inputTensorInfo.GetShape(), NCHWToNHWC, inputData.data(), tmp.data()); @@ -166,7 +167,7 @@ LayerTestResult SimpleConvolution2dTestImpl( outputData.insert(outputData.end(), outputImage.begin(), outputImage.end()); // at this point if we require it permute the expected output - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { std::vector tmp(outputData.size()); armnnUtils::Permute(outputTensorInfo.GetShape(), NCHWToNHWC, outputData.data(), tmp.data()); @@ -187,7 +188,7 @@ LayerTestResult SimpleConvolution2dTestImpl( armnn::ScopedCpuTensorHandle biasTensor(biasDesc); // Permute the kernel if necessary boost::multi_array kernel = boost::multi_array(originalKernel); - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { armnnUtils::Permute(kernelDesc.GetShape(), NCHWToNHWC, originalKernel.data(), kernel.data()); } @@ -210,7 +211,7 @@ LayerTestResult SimpleConvolution2dTestImpl( data.m_Parameters.m_PadTop = padTop; data.m_Parameters.m_PadBottom = padBottom; data.m_Parameters.m_BiasEnabled = biasEnabled; - data.m_Parameters.m_DataLayout = layout.GetDataLayout(); + data.m_Parameters.m_DataLayout = layout; std::unique_ptr workload = workloadFactory.CreateConvolution2d(data, info); inputHandle->Allocate(); @@ -327,7 +328,7 @@ LayerTestResult DepthwiseConvolution2dAsymmetricTestImpl( const boost::multi_array& outputExpected, float qScale, int32_t qOffset, - const armnn::DataLayoutIndexed& layout, + const armnn::DataLayout layout, uint32_t padLeft = 0, uint32_t padTop = 0, uint32_t padRight = 0, @@ -377,7 +378,7 @@ LayerTestResult DepthwiseConvolution2dAsymmetricTestImpl( // At this point if we require it permute the input data const armnn::PermutationVector NCHWToNHWC = { 0, 3, 1, 2 }; - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { std::vector tmp(inputData.size()); armnnUtils::Permute(inputTensorInfo.GetShape(), NCHWToNHWC, inputData.data(), tmp.data()); @@ -401,7 +402,7 @@ LayerTestResult DepthwiseConvolution2dAsymmetricTestImpl( LayerTestResult ret(outputTensorInfo); // At this point if we require it permute the expected output - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { std::vector tmp(outputData.size()); armnnUtils::Permute(outputTensorInfo.GetShape(), NCHWToNHWC, outputData.data(), tmp.data()); @@ -417,7 +418,7 @@ LayerTestResult DepthwiseConvolution2dAsymmetricTestImpl( // Permute the kernel if necessary boost::multi_array kernel = boost::multi_array(originalKernel); - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { armnnUtils::Permute(kernelDesc.GetShape(), NCHWToNHWC, originalKernel.data(), kernel.data()); } @@ -440,7 +441,7 @@ LayerTestResult DepthwiseConvolution2dAsymmetricTestImpl( data.m_Parameters.m_PadTop = padTop; data.m_Parameters.m_PadBottom = padBottom; data.m_Parameters.m_BiasEnabled = biasEnabled; - data.m_Parameters.m_DataLayout = layout.GetDataLayout(); + data.m_Parameters.m_DataLayout = layout; armnn::WorkloadInfo info; AddInputToWorkload(data, info, inputTensorInfo, inputHandle.get()); @@ -466,7 +467,7 @@ LayerTestResult DepthwiseConvolution2dDepthMul1TestImpl( float qScale, int32_t qOffset, bool biasEnabled, - const armnn::DataLayoutIndexed& layout) + const armnn::DataLayout layout) { unsigned int inputHeight = 3; unsigned int inputWidth = 3; @@ -511,7 +512,7 @@ LayerTestResult DepthwiseConvolution2dDepthMul1TestImpl( })); // at this point if we require it permute the input data const armnn::PermutationVector NCHWToNHWC = { 0, 3, 1, 2 }; - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { std::vector tmp(inputData.size()); armnnUtils::Permute(inputTensorInfo.GetShape(), NCHWToNHWC, inputData.data(), tmp.data()); @@ -533,7 +534,7 @@ LayerTestResult DepthwiseConvolution2dDepthMul1TestImpl( 0.f, 0.f, 0.f, -1.f, 0.f, -1.f, })); - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { std::vector tmp(kernelData.size()); armnnUtils::Permute(kernelDesc.GetShape(), NCHWToNHWC, kernelData.data(), tmp.data()); @@ -557,7 +558,7 @@ LayerTestResult DepthwiseConvolution2dDepthMul1TestImpl( } LayerTestResult ret(outputTensorInfo); - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { std::vector tmp(outputImage.size()); armnnUtils::Permute(outputTensorInfo.GetShape(), NCHWToNHWC, outputImage.data(), tmp.data()); @@ -589,7 +590,7 @@ LayerTestResult DepthwiseConvolution2dDepthMul1TestImpl( data.m_Parameters.m_PadTop = 0; data.m_Parameters.m_PadBottom = 0; data.m_Parameters.m_BiasEnabled = biasEnabled; - data.m_Parameters.m_DataLayout = layout.GetDataLayout(); + data.m_Parameters.m_DataLayout = layout; std::unique_ptr workload = workloadFactory.CreateDepthwiseConvolution2d(data, info); inputHandle->Allocate(); @@ -611,7 +612,7 @@ LayerTestResult DepthwiseConvolution2dTestImpl( float qScale, int32_t qOffset, bool biasEnabled, - const armnn::DataLayoutIndexed& layout) + const armnn::DataLayout layout) { unsigned int depthMultiplier = 2; @@ -672,7 +673,7 @@ LayerTestResult DepthwiseConvolution2dTestImpl( std::vector inputData = originalInputData; // at this point if we require it permute the input data const armnn::PermutationVector NCHWToNHWC = { 0, 3, 1, 2 }; - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { armnnUtils::Permute(inputTensorInfo.GetShape(), NCHWToNHWC, originalInputData.data(), inputData.data()); } @@ -709,7 +710,7 @@ LayerTestResult DepthwiseConvolution2dTestImpl( 0, 0, 0 })); std::vector kernelData = originalKernelData; - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { armnnUtils::Permute(kernelDesc.GetShape(), NCHWToNHWC, originalKernelData.data(), kernelData.data()); } @@ -762,7 +763,7 @@ LayerTestResult DepthwiseConvolution2dTestImpl( LayerTestResult ret(outputTensorInfo); std::vector outputImage = originalOutputImage; - if (layout.GetDataLayout() == armnn::DataLayout::NHWC) + if (layout == armnn::DataLayout::NHWC) { armnnUtils::Permute(outputTensorInfo.GetShape(), NCHWToNHWC, originalOutputImage.data(), outputImage.data()); } @@ -792,7 +793,7 @@ LayerTestResult DepthwiseConvolution2dTestImpl( data.m_Parameters.m_PadTop = 1; data.m_Parameters.m_PadBottom = 1; data.m_Parameters.m_BiasEnabled = biasEnabled; - data.m_Parameters.m_DataLayout = layout.GetDataLayout(); + data.m_Parameters.m_DataLayout = layout; std::unique_ptr workload = workloadFactory.CreateDepthwiseConvolution2d(data, info); inputHandle->Allocate(); -- cgit v1.2.1