// // Copyright © 2017 Arm Ltd. All rights reserved. // SPDX-License-Identifier: MIT // #pragma once #include "WorkloadTestUtils.hpp" #include #include #include #include #include #include #include namespace { template LayerTestResult DebugTestImpl( armnn::IWorkloadFactory& workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager, armnn::TensorInfo& inputTensorInfo, armnn::TensorInfo& outputTensorInfo, std::vector& inputData, std::vector& outputExpectedData, armnn::DebugQueueDescriptor descriptor, const std::string expectedStringOutput, const float qScale = 1.0f, const int32_t qOffset = 0) { if(armnn::IsQuantizedType()) { inputTensorInfo.SetQuantizationScale(qScale); inputTensorInfo.SetQuantizationOffset(qOffset); outputTensorInfo.SetQuantizationScale(qScale); outputTensorInfo.SetQuantizationOffset(qOffset); } boost::multi_array input = MakeTensor(inputTensorInfo, QuantizedVector(qScale, qOffset, inputData)); LayerTestResult ret(outputTensorInfo); ret.outputExpected = MakeTensor(outputTensorInfo, QuantizedVector(qScale, qOffset, outputExpectedData)); std::unique_ptr inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo); std::unique_ptr outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo); armnn::WorkloadInfo info; AddInputToWorkload(descriptor, info, inputTensorInfo, inputHandle.get()); AddOutputToWorkload(descriptor, info, outputTensorInfo, outputHandle.get()); std::unique_ptr workload = workloadFactory.CreateDebug(descriptor, info); inputHandle->Allocate(); outputHandle->Allocate(); CopyDataToITensorHandle(inputHandle.get(), input.data()); std::ostringstream oss; std::streambuf* coutStreambuf = std::cout.rdbuf(); std::cout.rdbuf(oss.rdbuf()); ExecuteWorkload(*workload, memoryManager); std::cout.rdbuf(coutStreambuf); BOOST_TEST(oss.str() == expectedStringOutput); CopyDataFromITensorHandle(ret.output.data(), outputHandle.get()); return ret; } template > LayerTestResult Debug4DTest( armnn::IWorkloadFactory& workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager) { armnn::TensorInfo inputTensorInfo; armnn::TensorInfo outputTensorInfo; unsigned int inputShape[] = {1, 2, 2, 3}; unsigned int outputShape[] = {1, 2, 2, 3}; armnn::DebugQueueDescriptor desc; desc.m_Guid = 1; desc.m_LayerName = "TestOutput"; desc.m_SlotIndex = 0; inputTensorInfo = armnn::TensorInfo(4, inputShape, ArmnnType); outputTensorInfo = armnn::TensorInfo(4, outputShape, ArmnnType); std::vector input = std::vector( { 1.0f, 2.0f, 3.0f, 4.0f, 5.0f, 6.0f, 7.0f, 8.0f, 9.0f, 10.0f, 11.0f, 12.0f, }); std::vector outputExpected = std::vector( { 1.0f, 2.0f, 3.0f, 4.0f, 5.0f, 6.0f, 7.0f, 8.0f, 9.0f, 10.0f, 11.0f, 12.0f, }); const std::string expectedStringOutput = "{ \"layerGuid\": 1," " \"layerName\": \"TestOutput\"," " \"outputSlot\": 0," " \"shape\": [1, 2, 2, 3]," " \"min\": 1, \"max\": 12," " \"data\": [[[[1, 2, 3], [4, 5, 6]], [[7, 8, 9], [10, 11, 12]]]] }\n"; return DebugTestImpl(workloadFactory, memoryManager, inputTensorInfo, outputTensorInfo, input, outputExpected, desc, expectedStringOutput); } template > LayerTestResult Debug3DTest( armnn::IWorkloadFactory& workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager) { armnn::TensorInfo inputTensorInfo; armnn::TensorInfo outputTensorInfo; unsigned int inputShape[] = {3, 3, 1}; unsigned int outputShape[] = {3, 3, 1}; armnn::DebugQueueDescriptor desc; desc.m_Guid = 1; desc.m_LayerName = "TestOutput"; desc.m_SlotIndex = 0; inputTensorInfo = armnn::TensorInfo(3, inputShape, ArmnnType); outputTensorInfo = armnn::TensorInfo(3, outputShape, ArmnnType); std::vector input = std::vector( { 1.0f, 2.0f, 3.0f, 4.0f, 5.0f, 6.0f, 7.0f, 8.0f, 9.0f, }); std::vector outputExpected = std::vector( { 1.0f, 2.0f, 3.0f, 4.0f, 5.0f, 6.0f, 7.0f, 8.0f, 9.0f, }); const std::string expectedStringOutput = "{ \"layerGuid\": 1," " \"layerName\": \"TestOutput\"," " \"outputSlot\": 0," " \"shape\": [3, 3, 1]," " \"min\": 1, \"max\": 9," " \"data\": [[[1], [2], [3]], [[4], [5], [6]], [[7], [8], [9]]] }\n"; return DebugTestImpl(workloadFactory, memoryManager, inputTensorInfo, outputTensorInfo, input, outputExpected, desc, expectedStringOutput); } template > LayerTestResult Debug2DTest( armnn::IWorkloadFactory& workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager) { armnn::TensorInfo inputTensorInfo; armnn::TensorInfo outputTensorInfo; unsigned int inputShape[] = {2, 2}; unsigned int outputShape[] = {2, 2}; armnn::DebugQueueDescriptor desc; desc.m_Guid = 1; desc.m_LayerName = "TestOutput"; desc.m_SlotIndex = 0; inputTensorInfo = armnn::TensorInfo(2, inputShape, ArmnnType); outputTensorInfo = armnn::TensorInfo(2, outputShape, ArmnnType); std::vector input = std::vector( { 1.0f, 2.0f, 3.0f, 4.0f, }); std::vector outputExpected = std::vector( { 1.0f, 2.0f, 3.0f, 4.0f, }); const std::string expectedStringOutput = "{ \"layerGuid\": 1," " \"layerName\": \"TestOutput\"," " \"outputSlot\": 0," " \"shape\": [2, 2]," " \"min\": 1, \"max\": 4," " \"data\": [[1, 2], [3, 4]] }\n"; return DebugTestImpl(workloadFactory, memoryManager, inputTensorInfo, outputTensorInfo, input, outputExpected, desc, expectedStringOutput); } template > LayerTestResult Debug1DTest( armnn::IWorkloadFactory& workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager) { armnn::TensorInfo inputTensorInfo; armnn::TensorInfo outputTensorInfo; unsigned int inputShape[] = {4}; unsigned int outputShape[] = {4}; armnn::DebugQueueDescriptor desc; desc.m_Guid = 1; desc.m_LayerName = "TestOutput"; desc.m_SlotIndex = 0; inputTensorInfo = armnn::TensorInfo(1, inputShape, ArmnnType); outputTensorInfo = armnn::TensorInfo(1, outputShape, ArmnnType); std::vector input = std::vector( { 1.0f, 2.0f, 3.0f, 4.0f, }); std::vector outputExpected = std::vector( { 1.0f, 2.0f, 3.0f, 4.0f, }); const std::string expectedStringOutput = "{ \"layerGuid\": 1," " \"layerName\": \"TestOutput\"," " \"outputSlot\": 0," " \"shape\": [4]," " \"min\": 1, \"max\": 4," " \"data\": [1, 2, 3, 4] }\n"; return DebugTestImpl(workloadFactory, memoryManager, inputTensorInfo, outputTensorInfo, input, outputExpected, desc, expectedStringOutput); } } // anonymous namespace