// // Copyright © 2017 Arm Ltd. All rights reserved. // SPDX-License-Identifier: MIT // #include #include #include #include namespace { template void CheckInputOutput(std::unique_ptr workload, const TensorInfo& inputInfo, const TensorInfo& outputInfo) { auto queueDescriptor = workload->GetData(); auto inputHandle = boost::polymorphic_downcast(queueDescriptor.m_Inputs[0]); auto outputHandle = boost::polymorphic_downcast(queueDescriptor.m_Outputs[0]); BOOST_TEST((inputHandle->GetTensorInfo() == inputInfo)); BOOST_TEST((outputHandle->GetTensorInfo() == outputInfo)); } template void CheckInputsOutput(std::unique_ptr workload, const TensorInfo& inputInfo0, const TensorInfo& inputInfo1, const TensorInfo& outputInfo) { auto queueDescriptor = workload->GetData(); auto inputHandle0 = boost::polymorphic_downcast(queueDescriptor.m_Inputs[0]); auto inputHandle1 = boost::polymorphic_downcast(queueDescriptor.m_Inputs[1]); auto outputHandle = boost::polymorphic_downcast(queueDescriptor.m_Outputs[0]); BOOST_TEST((inputHandle0->GetTensorInfo() == inputInfo0)); BOOST_TEST((inputHandle1->GetTensorInfo() == inputInfo1)); BOOST_TEST((outputHandle->GetTensorInfo() == outputInfo)); } } BOOST_AUTO_TEST_SUITE(CreateWorkloadRef) template static void RefCreateActivationWorkloadTest() { Graph graph; RefWorkloadFactory factory; auto workload = CreateActivationWorkloadTest(factory, graph); // Checks that outputs are as we expect them (see definition of CreateActivationWorkloadTest). CheckInputOutput(std::move(workload), TensorInfo({ 1, 1 }, DataType), TensorInfo({ 1, 1 }, DataType)); } BOOST_AUTO_TEST_CASE(CreateActivationFloat32Workload) { RefCreateActivationWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateActivationUint8Workload) { RefCreateActivationWorkloadTest(); } template static void RefCreateArithmethicWorkloadTest() { Graph graph; RefWorkloadFactory factory; auto workload = CreateArithmeticWorkloadTest(factory, graph); CheckInputsOutput(std::move(workload), TensorInfo({ 2, 3 }, DataType), TensorInfo({ 2, 3 }, DataType), TensorInfo({ 2, 3 }, DataType)); } BOOST_AUTO_TEST_CASE(CreateAdditionFloatWorkload) { RefCreateArithmethicWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateAdditionUint8Workload) { RefCreateArithmethicWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateSubtractionFloatWorkload) { RefCreateArithmethicWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateSubtractionUint8Workload) { RefCreateArithmethicWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateMultiplicationFloatWorkload) { RefCreateArithmethicWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateMultiplicationUint8Workload) { RefCreateArithmethicWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateDivisionFloatWorkload) { RefCreateArithmethicWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateDivisionUint8Workload) { RefCreateArithmethicWorkloadTest(); } template static void RefCreateBatchNormalizationWorkloadTest(DataLayout dataLayout) { Graph graph; RefWorkloadFactory factory; auto workload = CreateBatchNormalizationWorkloadTest(factory, graph, dataLayout); TensorShape inputShape; TensorShape outputShape; switch (dataLayout) { case DataLayout::NHWC: inputShape = { 2, 4, 4, 3 }; outputShape = { 2, 4, 4, 3 }; break; case DataLayout::NCHW: default: inputShape = { 2, 3, 4, 4 }; outputShape = { 2, 3, 4, 4 }; break; } // Checks that outputs and inputs are as we expect them (see definition of CreateBatchNormalizationWorkloadTest). CheckInputOutput(std::move(workload), TensorInfo(inputShape, DataType), TensorInfo(outputShape, DataType)); } BOOST_AUTO_TEST_CASE(CreateBatchNormalizationFloat32Workload) { RefCreateBatchNormalizationWorkloadTest (DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateBatchNormalizationFloat32WorkloadNhwc) { RefCreateBatchNormalizationWorkloadTest (DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(CreateBatchNormalizationUint8Workload) { RefCreateBatchNormalizationWorkloadTest (DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateBatchNormalizationUint8WorkloadNhwc) { RefCreateBatchNormalizationWorkloadTest (DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(CreateConvertFp16ToFp32Float32Workload) { Graph graph; RefWorkloadFactory factory; auto workload = CreateConvertFp16ToFp32WorkloadTest(factory, graph); // Checks that outputs and inputs are as we expect them CheckInputOutput( std::move(workload), TensorInfo({1, 3, 2, 3}, DataType::Float16), TensorInfo({1, 3, 2, 3}, DataType::Float32)); } BOOST_AUTO_TEST_CASE(CreateConvertFp32ToFp16Float16Workload) { Graph graph; RefWorkloadFactory factory; auto workload = CreateConvertFp32ToFp16WorkloadTest(factory, graph); // Checks that outputs and inputs are as we expect them CheckInputOutput( std::move(workload), TensorInfo({1, 3, 2, 3}, DataType::Float32), TensorInfo({1, 3, 2, 3}, DataType::Float16)); } static void RefCreateConvolution2dWorkloadTest(DataLayout dataLayout = DataLayout::NCHW) { Graph graph; RefWorkloadFactory factory; auto workload = CreateConvolution2dWorkloadTest (factory, graph, dataLayout); std::initializer_list inputShape = (dataLayout == DataLayout::NCHW) ? std::initializer_list({2, 3, 8, 16}) : std::initializer_list({2, 8, 16, 3}); std::initializer_list outputShape = (dataLayout == DataLayout::NCHW) ? std::initializer_list({2, 2, 2, 10}) : std::initializer_list({2, 2, 10, 2}); // Checks that outputs and inputs are as we expect them (see definition of CreateConvolution2dWorkloadTest). CheckInputOutput(std::move(workload), TensorInfo(inputShape, DataType::Float32), TensorInfo(outputShape, DataType::Float32)); } BOOST_AUTO_TEST_CASE(CreateConvolution2dFloatNchwWorkload) { RefCreateConvolution2dWorkloadTest(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateConvolution2dFloatNhwcWorkload) { RefCreateConvolution2dWorkloadTest(DataLayout::NHWC); } template static void RefCreateFullyConnectedWorkloadTest() { Graph graph; RefWorkloadFactory factory; auto workload = CreateFullyConnectedWorkloadTest(factory, graph); // Checks that outputs and inputs are as we expect them (see definition of CreateFullyConnectedWorkloadTest). float inputsQScale = DataType == armnn::DataType::QuantisedAsymm8 ? 1.0f : 0.0; float outputQScale = DataType == armnn::DataType::QuantisedAsymm8 ? 2.0f : 0.0; CheckInputOutput(std::move(workload), TensorInfo({ 3, 1, 4, 5 }, DataType, inputsQScale), TensorInfo({ 3, 7 }, DataType, outputQScale)); } BOOST_AUTO_TEST_CASE(CreateFullyConnectedFloat32Workload) { RefCreateFullyConnectedWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateFullyConnectedUint8Workload) { RefCreateFullyConnectedWorkloadTest(); } template static void RefCreateNormalizationWorkloadTest(DataLayout dataLayout) { Graph graph; RefWorkloadFactory factory; auto workload = CreateNormalizationWorkloadTest(factory, graph, dataLayout); TensorShape inputShape; TensorShape outputShape; switch (dataLayout) { case DataLayout::NHWC: inputShape = { 3, 1, 5, 5 }; outputShape = { 3, 1, 5, 5 }; break; case DataLayout::NCHW: default: inputShape = { 3, 5, 5, 1 }; outputShape = { 3, 5, 5, 1 }; break; } // Checks that outputs and inputs are as we expect them (see definition of CreateNormalizationWorkloadTest). CheckInputOutput(std::move(workload), TensorInfo(inputShape, DataType), TensorInfo(outputShape, DataType)); } BOOST_AUTO_TEST_CASE(CreateRefNormalizationNchwWorkload) { RefCreateNormalizationWorkloadTest(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateRefNormalizationNhwcWorkload) { RefCreateNormalizationWorkloadTest(DataLayout::NHWC); } template static void RefCreatePooling2dWorkloadTest() { Graph graph; RefWorkloadFactory factory; auto workload = CreatePooling2dWorkloadTest(factory, graph); // Checks that outputs and inputs are as we expect them (see definition of CreatePooling2dWorkloadTest). CheckInputOutput( std::move(workload), TensorInfo({3, 2, 5, 5}, DataType), TensorInfo({3, 2, 2, 4}, DataType)); } BOOST_AUTO_TEST_CASE(CreatePooling2dFloat32Workload) { RefCreatePooling2dWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreatePooling2dUint8Workload) { RefCreatePooling2dWorkloadTest(); } template static void RefCreateSoftmaxWorkloadTest() { Graph graph; RefWorkloadFactory factory; auto workload = CreateSoftmaxWorkloadTest(factory, graph); // Checks that outputs and inputs are as we expect them (see definition of CreateSoftmaxWorkloadTest). CheckInputOutput( std::move(workload), TensorInfo({4, 1}, DataType), TensorInfo({4, 1}, DataType)); } BOOST_AUTO_TEST_CASE(CreateSoftmaxFloat32Workload) { RefCreateSoftmaxWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateSoftmaxUint8Workload) { RefCreateSoftmaxWorkloadTest(); } template static void RefCreateSplitterWorkloadTest() { Graph graph; RefWorkloadFactory factory; auto workload = CreateSplitterWorkloadTest(factory, graph); // Checks that outputs are as we expect them (see definition of CreateSplitterWorkloadTest). SplitterQueueDescriptor queueDescriptor = workload->GetData(); auto inputHandle = boost::polymorphic_downcast(queueDescriptor.m_Inputs[0]); BOOST_TEST((inputHandle->GetTensorInfo() == TensorInfo({ 5, 7, 7 }, DataType))); auto outputHandle0 = boost::polymorphic_downcast(queueDescriptor.m_Outputs[0]); BOOST_TEST((outputHandle0->GetTensorInfo() == TensorInfo({ 1, 7, 7 }, DataType))); auto outputHandle1 = boost::polymorphic_downcast(queueDescriptor.m_Outputs[1]); BOOST_TEST((outputHandle1->GetTensorInfo() == TensorInfo({ 2, 7, 7 }, DataType))); auto outputHandle2 = boost::polymorphic_downcast(queueDescriptor.m_Outputs[2]); BOOST_TEST((outputHandle2->GetTensorInfo() == TensorInfo({ 2, 7, 7 }, DataType))); } BOOST_AUTO_TEST_CASE(CreateSplitterFloat32Workload) { RefCreateSplitterWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateSplitterUint8Workload) { RefCreateSplitterWorkloadTest(); } template static void RefCreateSplitterMergerWorkloadTest() { // Tests that it is possible to decide which output of the splitter layer // should be lined to which input of the merger layer. // We tested that is is possible to specify 0th output // of the splitter to be the 1st input to the merger and the 1st output of the splitter to be 0th input // of the merger. Graph graph; RefWorkloadFactory factory; auto workloads = CreateSplitterMergerWorkloadTest (factory, graph); auto wlSplitter = std::move(workloads.first); auto wlMerger = std::move(workloads.second); //Checks that the index of inputs/outputs matches what we declared on InputDescriptor construction. armnn::CpuTensorHandle* sOut0 = dynamic_cast(wlSplitter->GetData().m_Outputs[0]); armnn::CpuTensorHandle* sOut1 = dynamic_cast(wlSplitter->GetData().m_Outputs[1]); armnn::CpuTensorHandle* mIn0 = dynamic_cast(wlMerger->GetData().m_Inputs[0]); armnn::CpuTensorHandle* mIn1 = dynamic_cast(wlMerger->GetData().m_Inputs[1]); BOOST_TEST(sOut0); BOOST_TEST(sOut1); BOOST_TEST(mIn0); BOOST_TEST(mIn1); bool validDataPointers = (sOut0 == mIn1) && (sOut1 == mIn0); BOOST_TEST(validDataPointers); } BOOST_AUTO_TEST_CASE(CreateSplitterMergerFloat32) { RefCreateSplitterMergerWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateSplitterMergerUint8) { RefCreateSplitterMergerWorkloadTest(); } template static void RefCreateSingleOutputMultipleInputsTest() { // Tests that it is possible to assign multiple (two) different layers to each of the outputs of a splitter layer. // We created a splitter with two outputs. That each of those outputs is used by two different activation layers. Graph graph; RefWorkloadFactory factory; std::unique_ptr wlSplitter; std::unique_ptr wlActiv0_0; std::unique_ptr wlActiv0_1; std::unique_ptr wlActiv1_0; std::unique_ptr wlActiv1_1; CreateSplitterMultipleInputsOneOutputWorkloadTest(factory, graph, wlSplitter, wlActiv0_0, wlActiv0_1, wlActiv1_0, wlActiv1_1); armnn::CpuTensorHandle* sOut0 = dynamic_cast(wlSplitter->GetData().m_Outputs[0]); armnn::CpuTensorHandle* sOut1 = dynamic_cast(wlSplitter->GetData().m_Outputs[1]); armnn::CpuTensorHandle* activ0_0Im = dynamic_cast(wlActiv0_0->GetData().m_Inputs[0]); armnn::CpuTensorHandle* activ0_1Im = dynamic_cast(wlActiv0_1->GetData().m_Inputs[0]); armnn::CpuTensorHandle* activ1_0Im = dynamic_cast(wlActiv1_0->GetData().m_Inputs[0]); armnn::CpuTensorHandle* activ1_1Im = dynamic_cast(wlActiv1_1->GetData().m_Inputs[0]); BOOST_TEST(sOut0); BOOST_TEST(sOut1); BOOST_TEST(activ0_0Im); BOOST_TEST(activ0_1Im); BOOST_TEST(activ1_0Im); BOOST_TEST(activ1_1Im); bool validDataPointers = (sOut0 == activ0_0Im) && (sOut0 == activ0_1Im) && (sOut1 == activ1_0Im) && (sOut1 == activ1_1Im); BOOST_TEST(validDataPointers); } BOOST_AUTO_TEST_CASE(CreateSingleOutputMultipleInputsFloat32) { RefCreateSingleOutputMultipleInputsTest(); } BOOST_AUTO_TEST_CASE(CreateSingleOutputMultipleInputsUint8) { RefCreateSingleOutputMultipleInputsTest(); } template static void RefCreateResizeBilinearTest(DataLayout dataLayout) { Graph graph; RefWorkloadFactory factory; auto workload = CreateResizeBilinearWorkloadTest(factory, graph, dataLayout); TensorShape inputShape; TensorShape outputShape; switch (dataLayout) { case DataLayout::NHWC: inputShape = { 2, 4, 4, 3 }; outputShape = { 2, 2, 2, 3 }; break; default: // NCHW inputShape = { 2, 3, 4, 4 }; outputShape = { 2, 3, 2, 2 }; } // Checks that outputs and inputs are as we expect them (see definition of CreateResizeBilinearWorkloadTest). CheckInputOutput( std::move(workload), TensorInfo(inputShape, DataType), TensorInfo(outputShape, DataType)); } BOOST_AUTO_TEST_CASE(CreateResizeBilinearFloat32) { RefCreateResizeBilinearTest(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateResizeBilinearUint8) { RefCreateResizeBilinearTest(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateResizeBilinearFloat32Nhwc) { RefCreateResizeBilinearTest(DataLayout::NHWC); } template static void RefCreateL2NormalizationTest(DataLayout dataLayout) { Graph graph; RefWorkloadFactory factory; auto workload = CreateL2NormalizationWorkloadTest(factory, graph, dataLayout); TensorShape inputShape; TensorShape outputShape; switch (dataLayout) { case DataLayout::NHWC: inputShape = { 5, 50, 67, 20 }; outputShape = { 5, 50, 67, 20 }; break; case DataLayout::NCHW: default: inputShape = { 5, 20, 50, 67 }; outputShape = { 5, 20, 50, 67 }; break; } // Checks that outputs and inputs are as we expect them (see definition of CreateL2NormalizationWorkloadTest). CheckInputOutput(std::move(workload), TensorInfo(inputShape, DataType), TensorInfo(outputShape, DataType)); } BOOST_AUTO_TEST_CASE(CreateL2NormalizationFloat32) { RefCreateL2NormalizationTest(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateL2NormalizationFloat32Nhwc) { RefCreateL2NormalizationTest(DataLayout::NHWC); } template static void RefCreateReshapeWorkloadTest() { Graph graph; RefWorkloadFactory factory; auto workload = CreateReshapeWorkloadTest(factory, graph); // Checks that outputs and inputs are as we expect them (see definition of CreateReshapeWorkloadTest). CheckInputOutput( std::move(workload), TensorInfo({ 4, 1 }, DataType), TensorInfo({ 1, 4 }, DataType)); } BOOST_AUTO_TEST_CASE(CreateReshapeFloat32Workload) { RefCreateReshapeWorkloadTest(); } BOOST_AUTO_TEST_CASE(CreateReshapeUint8Workload) { RefCreateReshapeWorkloadTest(); } BOOST_AUTO_TEST_SUITE_END()