// // Copyright © 2017 Arm Ltd. All rights reserved. // SPDX-License-Identifier: MIT // #include #include #include #include #include #include #include #include #include #include #include BOOST_AUTO_TEST_SUITE(NeonEndToEnd) std::vector defaultBackends = {armnn::Compute::CpuAcc}; // Abs BOOST_AUTO_TEST_CASE(NeonAbsEndToEndTestFloat32) { AbsEndToEnd(defaultBackends); } // Constant BOOST_AUTO_TEST_CASE(ConstantUsage_Neon_Float32) { BOOST_TEST(ConstantUsageFloat32Test(defaultBackends)); } #if defined(ARMNNREF_ENABLED) // This test unit needs the reference backend, it's not available if the reference backend is not built BOOST_AUTO_TEST_CASE(FallbackToCpuRef) { using namespace armnn; // Create runtime in which test will run and allow fallback to CpuRef. IRuntime::CreationOptions options; IRuntimePtr runtime(IRuntime::Create(options)); // Builds up the structure of the network. INetworkPtr net(INetwork::Create()); IConnectableLayer* input = net->AddInputLayer(0); // This layer configuration isn't supported by CpuAcc but we allow fallback to CpuRef so it shoud pass. NormalizationDescriptor descriptor; IConnectableLayer* pooling = net->AddNormalizationLayer(descriptor); IConnectableLayer* output = net->AddOutputLayer(0); input->GetOutputSlot(0).Connect(pooling->GetInputSlot(0)); pooling->GetOutputSlot(0).Connect(output->GetInputSlot(0)); input->GetOutputSlot(0).SetTensorInfo(TensorInfo({ 1, 1, 4, 4 }, DataType::Float32)); pooling->GetOutputSlot(0).SetTensorInfo(TensorInfo({ 1, 1, 4, 4 }, DataType::Float32)); // optimize the network std::vector backends = {Compute::CpuAcc, Compute::CpuRef}; IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec()); // Load it into the runtime. It should pass. NetworkId netId; BOOST_TEST(runtime->LoadNetwork(netId, std::move(optNet)) == Status::Success); } #endif BOOST_AUTO_TEST_CASE(NeonGreaterSimpleEndToEndTest) { const std::vector expectedOutput({ 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0 }); ArithmeticSimpleEndToEnd(defaultBackends, LayerType::Greater, expectedOutput); } BOOST_AUTO_TEST_CASE(NeonGreaterSimpleEndToEndUint8Test) { const std::vector expectedOutput({ 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0 }); ArithmeticSimpleEndToEnd(defaultBackends, LayerType::Greater, expectedOutput); } BOOST_AUTO_TEST_CASE(NeonGreaterBroadcastEndToEndTest) { const std::vector expectedOutput({ 0, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1 }); ArithmeticBroadcastEndToEnd(defaultBackends, LayerType::Greater, expectedOutput); } BOOST_AUTO_TEST_CASE(NeonGreaterBroadcastEndToEndUint8Test) { const std::vector expectedOutput({ 0, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1 }); ArithmeticBroadcastEndToEnd(defaultBackends, LayerType::Greater, expectedOutput); } BOOST_AUTO_TEST_CASE(NeonConcatEndToEndDim0Test) { ConcatDim0EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonConcatEndToEndDim0Uint8Test) { ConcatDim0EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonConcatEndToEndDim1Test) { ConcatDim1EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonConcatEndToEndDim1Uint8Test) { ConcatDim1EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonConcatEndToEndDim3Test) { ConcatDim3EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonConcatEndToEndDim3Uint8Test) { ConcatDim3EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(DequantizeEndToEndSimpleTest) { DequantizeEndToEndSimple(defaultBackends); } BOOST_AUTO_TEST_CASE(DequantizeEndToEndOffsetTest) { DequantizeEndToEndOffset(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonPreluEndToEndFloat32Test) { PreluEndToEndNegativeTest(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonPreluEndToEndTestUint8Test) { PreluEndToEndPositiveTest(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSpaceToDepthNHWCEndToEndTest1) { SpaceToDepthNHWCEndToEndTest1(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSpaceToDepthNCHWEndToEndTest1) { SpaceToDepthNCHWEndToEndTest1(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSpaceToDepthNHWCEndToEndTest2) { SpaceToDepthNHWCEndToEndTest2(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSpaceToDepthNCHWEndToEndTest2) { SpaceToDepthNCHWEndToEndTest2(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter1dEndToEndTest) { Splitter1dEndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter1dEndToEndUint8Test) { Splitter1dEndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter2dDim0EndToEndTest) { Splitter2dDim0EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter2dDim1EndToEndTest) { Splitter2dDim1EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter2dDim0EndToEndUint8Test) { Splitter2dDim0EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter2dDim1EndToEndUint8Test) { Splitter2dDim1EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter3dDim0EndToEndTest) { Splitter3dDim0EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter3dDim1EndToEndTest) { Splitter3dDim1EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter3dDim2EndToEndTest) { Splitter3dDim2EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter3dDim0EndToEndUint8Test) { Splitter3dDim0EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter3dDim1EndToEndUint8Test) { Splitter3dDim1EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter3dDim2EndToEndUint8Test) { Splitter3dDim2EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter4dDim0EndToEndTest) { Splitter4dDim0EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter4dDim1EndToEndTest) { Splitter4dDim1EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter4dDim2EndToEndTest) { Splitter4dDim2EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter4dDim3EndToEndTest) { Splitter4dDim3EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter4dDim0EndToEndUint8Test) { Splitter4dDim0EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter4dDim1EndToEndUint8Test) { Splitter4dDim1EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter4dDim2EndToEndUint8Test) { Splitter4dDim2EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonSplitter4dDim3EndToEndUint8Test) { Splitter4dDim3EndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonQuantizedLstmEndToEndTest) { QuantizedLstmEndToEnd(defaultBackends); } BOOST_AUTO_TEST_CASE(NeonTransposeConvolution2dEndToEndFloatNchwTest) { TransposeConvolution2dEndToEnd( defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(NeonTransposeConvolution2dEndToEndUint8NchwTest) { TransposeConvolution2dEndToEnd( defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(NeonTransposeConvolution2dEndToEndFloatNhwcTest) { TransposeConvolution2dEndToEnd( defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(NeonTransposeConvolution2dEndToEndUint8NhwcTest) { TransposeConvolution2dEndToEnd( defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(NeonImportNonAlignedInputPointerTest) { ImportNonAlignedInputPointerTest(defaultBackends); } // Utility function to find the number of instances of a substring within a string. int SubStringCounter(std::string& string, std::string&& substring) { std::size_t found = 0; int count = 0; // Look for the substring starting from where we last found the substring while((found = string.find(substring, found)) != std::string::npos) { count++; // Offset by substring length to avoid finding the same substring twice found += substring.length(); } return count; } BOOST_AUTO_TEST_CASE(NeonImportOnlyWorkload) { using namespace armnn; IRuntime::CreationOptions options; IRuntimePtr runtime(IRuntime::Create(options)); // Builds up the structure of the network. INetworkPtr net(INetwork::Create()); IConnectableLayer* input = net->AddInputLayer(0); ActivationDescriptor descriptor; descriptor.m_Function = ActivationFunction::Square; IConnectableLayer* pooling = net->AddActivationLayer(descriptor); IConnectableLayer* output = net->AddOutputLayer(0); input->GetOutputSlot(0).Connect(pooling->GetInputSlot(0)); pooling->GetOutputSlot(0).Connect(output->GetInputSlot(0)); input->GetOutputSlot(0).SetTensorInfo(TensorInfo({ 1, 1, 1, 4 }, DataType::Float32)); pooling->GetOutputSlot(0).SetTensorInfo(TensorInfo({ 1, 1, 1, 4 }, DataType::Float32)); // optimize the network std::vector backends = {Compute::CpuAcc}; IOptimizedNetworkPtr optNet = Optimize(*net, backends, runtime->GetDeviceSpec()); BOOST_TEST_CHECKPOINT("Load Network"); // Load it into the runtime. It should pass. NetworkId netId; std::string ignoredErrorMessage; INetworkProperties networkProperties(true, false); BOOST_TEST(runtime->LoadNetwork(netId, std::move(optNet),ignoredErrorMessage, networkProperties) == Status::Success); BOOST_TEST_CHECKPOINT("Generate Data"); // Creates structures for input & output std::vector inputData { 1.0f, 2.0f, 3.0f, 4.0f }; std::vector outputData(4); std::vector expectedOutput { 1.0f, 4.0f, 9.0f, 16.0f }; BOOST_TEST_CHECKPOINT("Create Network"); InputTensors inputTensors { {0,armnn::ConstTensor(runtime->GetInputTensorInfo(netId, 0), inputData.data())}, }; OutputTensors outputTensors { {0,armnn::Tensor(runtime->GetOutputTensorInfo(netId, 0), outputData.data())} }; BOOST_TEST_CHECKPOINT("Get Profiler"); runtime->GetProfiler(netId)->EnableProfiling(true); BOOST_TEST_CHECKPOINT("Run Inference"); // Do the inference runtime->EnqueueWorkload(netId, inputTensors, outputTensors); BOOST_TEST_CHECKPOINT("Print Profiler"); // Retrieve the Profiler.Print() output to get the workload execution ProfilerManager& profilerManager = armnn::ProfilerManager::GetInstance(); std::stringstream ss; profilerManager.GetProfiler()->Print(ss);; std::string dump = ss.str(); // Check there are no SyncMemGeneric workloads as we didn't export BOOST_TEST_CHECKPOINT("Find SyncMemGeneric"); int count = SubStringCounter(dump, "SyncMemGeneric"); BOOST_TEST(count == 0); // Should only be 1 CopyMemGeneric for the output as we imported BOOST_TEST_CHECKPOINT("Find CopyMemGeneric"); count = SubStringCounter(dump, "CopyMemGeneric"); BOOST_TEST(count == 1); // Check the output is correct BOOST_CHECK_EQUAL_COLLECTIONS(outputData.begin(), outputData.end(), expectedOutput.begin(), expectedOutput.end()); } BOOST_AUTO_TEST_SUITE_END()