diff options
Diffstat (limited to 'src/armnn/test/Network_test.cpp')
-rw-r--r-- | src/armnn/test/Network_test.cpp | 483 |
1 files changed, 0 insertions, 483 deletions
diff --git a/src/armnn/test/Network_test.cpp b/src/armnn/test/Network_test.cpp deleted file mode 100644 index 057caa0505..0000000000 --- a/src/armnn/test/Network_test.cpp +++ /dev/null @@ -1,483 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// See LICENSE file in the project root for full license information. -// -#include <boost/test/unit_test.hpp> - -#include "armnn/ArmNN.hpp" -#include "Network.hpp" -#include "Graph.hpp" -#include "backends/RefWorkloadFactory.hpp" - -#include "GraphUtils.hpp" - -namespace -{ - -bool AreAllLayerInputSlotsConnected(const armnn::IConnectableLayer& layer) -{ - bool allConnected = true; - for (unsigned int i = 0; i < layer.GetNumInputSlots(); ++i) - { - const bool inputConnected = layer.GetInputSlot(i).GetConnection() != nullptr; - allConnected &= inputConnected; - } - return allConnected; -} - -} - -BOOST_AUTO_TEST_SUITE(Network) - -BOOST_AUTO_TEST_CASE(LayerGuids) -{ - armnn::Network net; - armnn::LayerGuid inputId = net.AddInputLayer(0)->GetGuid(); - armnn::LayerGuid addId = net.AddAdditionLayer()->GetGuid(); - armnn::LayerGuid outputId = net.AddOutputLayer(0)->GetGuid(); - - BOOST_TEST(inputId != addId); - BOOST_TEST(addId != outputId); - BOOST_TEST(inputId != outputId); -} - -BOOST_AUTO_TEST_CASE(SerializeToDot) -{ - armnn::Network net; - - //define layers - auto input = net.AddInputLayer(0); - auto add = net.AddAdditionLayer(); - auto output = net.AddOutputLayer(0); - - // connect layers - input->GetOutputSlot(0).Connect(add->GetInputSlot(0)); - input->GetOutputSlot(0).Connect(add->GetInputSlot(1)); - add->GetOutputSlot(0).Connect(output->GetInputSlot(0)); - - armnn::TensorShape shape({4}); - armnn::TensorInfo info(shape, armnn::DataType::Float32); - input->GetOutputSlot(0).SetTensorInfo(info); - add->GetOutputSlot(0).SetTensorInfo(info); - - armnn::DeviceSpec spec; - spec.DefaultComputeDevice = armnn::Compute::CpuAcc; - armnn::IOptimizedNetworkPtr optimizedNet = armnn::Optimize(net, spec); - - std::ostringstream ss; - optimizedNet->SerializeToDot(ss); - - auto inputId = input->GetGuid(); - auto addId = add->GetGuid(); - auto outputId = output->GetGuid(); - - std::stringstream expected; - expected << - "digraph Optimized {\n" - " node [shape=\"record\"];\n" - " edge [fontsize=8 fontcolor=\"blue\" fontname=\"arial-bold\"];\n" - " " << inputId << " [label=\"{Input}\"];\n" - " " << addId << " [label=\"{Addition}\"];\n" - " " << outputId << " [label=\"{Output}\"];\n" - " " << inputId << " -> " << addId << " [label=< [4] >];\n" - " " << inputId << " -> " << addId << " [label=< [4] >];\n" - " " << addId << " -> " << outputId << " [label=< [4] >];\n" - "}\n"; - - BOOST_TEST(ss.str() == expected.str()); -} - -BOOST_AUTO_TEST_CASE(NetworkBasic) -{ - armnn::Network net; - BOOST_TEST(net.PrintGraph() == armnn::Status::Success); -} - -BOOST_AUTO_TEST_CASE(LayerNamesAreOptionalForINetwork) -{ - armnn::Network net; - armnn::INetwork& inet = net; - inet.AddInputLayer(0); - inet.AddAdditionLayer(); - inet.AddActivationLayer(armnn::ActivationDescriptor()); - inet.AddOutputLayer(0); -} - -BOOST_AUTO_TEST_CASE(LayerNamesAreOptionalForNetwork) -{ - armnn::Network net; - net.AddInputLayer(0); - net.AddAdditionLayer(); - net.AddActivationLayer(armnn::ActivationDescriptor()); - net.AddOutputLayer(0); -} - -BOOST_AUTO_TEST_CASE(NetworkModification) -{ - armnn::Network net; - - armnn::IConnectableLayer* const inputLayer = net.AddInputLayer(0, "input layer"); - BOOST_TEST(inputLayer); - - unsigned int dims[] = { 10,1,1,1 }; - std::vector<float> convWeightsData(10); - armnn::ConstTensor weights(armnn::TensorInfo(4, dims, armnn::DataType::Float32), convWeightsData); - - armnn::Convolution2dDescriptor convDesc2d; - armnn::IConnectableLayer* const convLayer = net.AddConvolution2dLayer(convDesc2d, weights, "conv layer"); - BOOST_TEST(convLayer); - - inputLayer->GetOutputSlot(0).Connect(convLayer->GetInputSlot(0)); - - armnn::FullyConnectedDescriptor fullyConnectedDesc; - armnn::IConnectableLayer* const fullyConnectedLayer = net.AddFullyConnectedLayer(fullyConnectedDesc, - weights, - "fully connected"); - BOOST_TEST(fullyConnectedLayer); - - convLayer->GetOutputSlot(0).Connect(fullyConnectedLayer->GetInputSlot(0)); - - armnn::Pooling2dDescriptor pooling2dDesc; - armnn::IConnectableLayer* const poolingLayer = net.AddPooling2dLayer(pooling2dDesc, "pooling2d"); - BOOST_TEST(poolingLayer); - - fullyConnectedLayer->GetOutputSlot(0).Connect(poolingLayer->GetInputSlot(0)); - - armnn::ActivationDescriptor activationDesc; - armnn::IConnectableLayer* const activationLayer = net.AddActivationLayer(activationDesc, "activation"); - BOOST_TEST(activationLayer); - - poolingLayer->GetOutputSlot(0).Connect(activationLayer->GetInputSlot(0)); - - armnn::NormalizationDescriptor normalizationDesc; - armnn::IConnectableLayer* const normalizationLayer = net.AddNormalizationLayer(normalizationDesc, "normalization"); - BOOST_TEST(normalizationLayer); - - activationLayer->GetOutputSlot(0).Connect(normalizationLayer->GetInputSlot(0)); - - armnn::SoftmaxDescriptor softmaxDesc; - armnn::IConnectableLayer* const softmaxLayer = net.AddSoftmaxLayer(softmaxDesc, "softmax"); - BOOST_TEST(softmaxLayer); - - normalizationLayer->GetOutputSlot(0).Connect(softmaxLayer->GetInputSlot(0)); - - armnn::BatchNormalizationDescriptor batchNormDesc; - - armnn::TensorInfo tensorInfo({ 1 }, armnn::DataType::Float32); - std::vector<float> data(tensorInfo.GetNumBytes() / sizeof(float)); - armnn::ConstTensor invalidTensor(tensorInfo, data); - - armnn::IConnectableLayer* const batchNormalizationLayer = net.AddBatchNormalizationLayer(batchNormDesc, - invalidTensor, - invalidTensor, - invalidTensor, - invalidTensor, - "batch norm"); - BOOST_TEST(batchNormalizationLayer); - - softmaxLayer->GetOutputSlot(0).Connect(batchNormalizationLayer->GetInputSlot(0)); - - armnn::IConnectableLayer* const additionLayer = net.AddAdditionLayer("addition"); - BOOST_TEST(additionLayer); - - batchNormalizationLayer->GetOutputSlot(0).Connect(additionLayer->GetInputSlot(0)); - batchNormalizationLayer->GetOutputSlot(0).Connect(additionLayer->GetInputSlot(1)); - - armnn::IConnectableLayer* const multiplicationLayer = net.AddMultiplicationLayer("multiplication"); - BOOST_TEST(multiplicationLayer); - - additionLayer->GetOutputSlot(0).Connect(multiplicationLayer->GetInputSlot(0)); - additionLayer->GetOutputSlot(0).Connect(multiplicationLayer->GetInputSlot(1)); - - armnn::IConnectableLayer* const outputLayer = net.AddOutputLayer(0, "output layer"); - BOOST_TEST(outputLayer); - - multiplicationLayer->GetOutputSlot(0).Connect(outputLayer->GetInputSlot(0)); - - //Test that all layers are present in the graph - BOOST_TEST(net.GetGraph().GetNumLayers() == 11); - - //Test that the vertices exist and have correct names - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "input layer")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "conv layer")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "fully connected")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "pooling2d")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "activation")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "normalization")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "softmax")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "batch norm")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "addition")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "multiplication")); - BOOST_TEST(GraphHasNamedLayer(net.GetGraph(), "output layer")); - - auto checkOneOutputToOneInputConnection = [] - (const armnn::IConnectableLayer* const srcLayer, - const armnn::IConnectableLayer* const tgtLayer, - int expectedSrcNumInputs = 1, - int expectedDstNumOutputs = 1) - { - BOOST_TEST(srcLayer->GetNumInputSlots() == expectedSrcNumInputs); - BOOST_TEST(srcLayer->GetNumOutputSlots() == 1); - BOOST_TEST(tgtLayer->GetNumInputSlots() == 1); - BOOST_TEST(tgtLayer->GetNumOutputSlots() == expectedDstNumOutputs); - - BOOST_TEST(srcLayer->GetOutputSlot(0).GetNumConnections() == 1); - BOOST_TEST(srcLayer->GetOutputSlot(0).GetConnection(0) == &tgtLayer->GetInputSlot(0)); - BOOST_TEST(&srcLayer->GetOutputSlot(0) == tgtLayer->GetInputSlot(0).GetConnection()); - }; - auto checkOneOutputToTwoInputsConnections = [] - (const armnn::IConnectableLayer* const srcLayer, - const armnn::IConnectableLayer* const tgtLayer, - int expectedSrcNumInputs, - int expectedDstNumOutputs = 1) - { - BOOST_TEST(srcLayer->GetNumInputSlots() == expectedSrcNumInputs); - BOOST_TEST(srcLayer->GetNumOutputSlots() == 1); - BOOST_TEST(tgtLayer->GetNumInputSlots() == 2); - BOOST_TEST(tgtLayer->GetNumOutputSlots() == expectedDstNumOutputs); - - BOOST_TEST(srcLayer->GetOutputSlot(0).GetNumConnections() == 2); - for (unsigned int i = 0; i < srcLayer->GetOutputSlot(0).GetNumConnections(); ++i) - { - BOOST_TEST(srcLayer->GetOutputSlot(0).GetConnection(i) == &tgtLayer->GetInputSlot(i)); - BOOST_TEST(&srcLayer->GetOutputSlot(0) == tgtLayer->GetInputSlot(i).GetConnection()); - } - }; - - BOOST_TEST(AreAllLayerInputSlotsConnected(*convLayer)); - BOOST_TEST(AreAllLayerInputSlotsConnected(*fullyConnectedLayer)); - BOOST_TEST(AreAllLayerInputSlotsConnected(*poolingLayer)); - BOOST_TEST(AreAllLayerInputSlotsConnected(*activationLayer)); - BOOST_TEST(AreAllLayerInputSlotsConnected(*normalizationLayer)); - BOOST_TEST(AreAllLayerInputSlotsConnected(*softmaxLayer)); - BOOST_TEST(AreAllLayerInputSlotsConnected(*batchNormalizationLayer)); - BOOST_TEST(AreAllLayerInputSlotsConnected(*additionLayer)); - BOOST_TEST(AreAllLayerInputSlotsConnected(*multiplicationLayer)); - BOOST_TEST(AreAllLayerInputSlotsConnected(*outputLayer)); - - // Check connectivity - checkOneOutputToOneInputConnection(inputLayer, convLayer, 0); - checkOneOutputToOneInputConnection(convLayer, fullyConnectedLayer); - checkOneOutputToOneInputConnection(fullyConnectedLayer, poolingLayer); - checkOneOutputToOneInputConnection(poolingLayer, activationLayer); - checkOneOutputToOneInputConnection(activationLayer, normalizationLayer); - checkOneOutputToOneInputConnection(normalizationLayer, softmaxLayer); - checkOneOutputToOneInputConnection(softmaxLayer, batchNormalizationLayer); - checkOneOutputToTwoInputsConnections(batchNormalizationLayer, additionLayer, 1); - checkOneOutputToTwoInputsConnections(additionLayer, multiplicationLayer, 2); - checkOneOutputToOneInputConnection(multiplicationLayer, outputLayer, 2, 0); -} - -BOOST_AUTO_TEST_CASE(NetworkModification_SplitterMerger) -{ - armnn::Network net; - - // Add an input layer and an input tensor descriptor. - armnn::IConnectableLayer* inputLayer = net.AddInputLayer(0, "input layer"); - BOOST_TEST(inputLayer); - - // Add a splitter layer - armnn::ViewsDescriptor splitterDesc(2,4); - - armnn::IConnectableLayer* splitterLayer = net.AddSplitterLayer(splitterDesc, "splitter layer"); - BOOST_TEST(splitterLayer); - - inputLayer->GetOutputSlot(0).Connect(splitterLayer->GetInputSlot(0)); - - // Add a softmax layer 1 - armnn::SoftmaxDescriptor softmaxDescriptor; - armnn::IConnectableLayer* softmaxLayer1 = net.AddSoftmaxLayer(softmaxDescriptor, "softmax_1"); - BOOST_TEST(softmaxLayer1); - - splitterLayer->GetOutputSlot(0).Connect(softmaxLayer1->GetInputSlot(0)); - - // Add a softmax layer 2 - armnn::IConnectableLayer* softmaxLayer2 = net.AddSoftmaxLayer(softmaxDescriptor, "softmax_2"); - BOOST_TEST(softmaxLayer2); - - splitterLayer->GetOutputSlot(1).Connect(softmaxLayer2->GetInputSlot(0)); - - // Add a merger layer - armnn::OriginsDescriptor mergerDesc(2, 4); - - armnn::IConnectableLayer* mergerLayer = net.AddMergerLayer(mergerDesc, "merger layer"); - BOOST_TEST(mergerLayer); - - softmaxLayer1->GetOutputSlot(0).Connect(mergerLayer->GetInputSlot(0)); - softmaxLayer2->GetOutputSlot(0).Connect(mergerLayer->GetInputSlot(1)); - - // Add an output layer - armnn::IConnectableLayer* outputLayer = net.AddOutputLayer(0, "output layer"); - BOOST_TEST(outputLayer); - - mergerLayer->GetOutputSlot(0).Connect(outputLayer->GetInputSlot(0)); - - BOOST_TEST(splitterLayer->GetNumOutputSlots() == 2); - BOOST_TEST(splitterLayer->GetOutputSlot(0).GetConnection(0) == &softmaxLayer1->GetInputSlot(0)); - BOOST_TEST(&splitterLayer->GetOutputSlot(0) == softmaxLayer1->GetInputSlot(0).GetConnection()); - BOOST_TEST(splitterLayer->GetOutputSlot(1).GetConnection(0) == &softmaxLayer2->GetInputSlot(0)); - BOOST_TEST(&splitterLayer->GetOutputSlot(1) == softmaxLayer2->GetInputSlot(0).GetConnection()); - - BOOST_TEST(mergerLayer->GetNumInputSlots() == 2); - BOOST_TEST(softmaxLayer1->GetOutputSlot(0).GetConnection(0) == &mergerLayer->GetInputSlot(0)); - BOOST_TEST(&softmaxLayer1->GetOutputSlot(0) == mergerLayer->GetInputSlot(0).GetConnection()); - BOOST_TEST(softmaxLayer2->GetOutputSlot(0).GetConnection(0) == &mergerLayer->GetInputSlot(1)); - BOOST_TEST(&softmaxLayer2->GetOutputSlot(0) == mergerLayer->GetInputSlot(1).GetConnection()); -} - -BOOST_AUTO_TEST_CASE(NetworkModification_SplitterAddition) -{ - armnn::Network net; - - // Add an input layer and an input tensor descriptor. - armnn::IConnectableLayer* layer = net.AddInputLayer(0, "input layer"); - BOOST_TEST(layer); - - // Add a splitter layer - armnn::ViewsDescriptor splitterDesc(2,4); - - armnn::IConnectableLayer* const splitterLayer = net.AddSplitterLayer(splitterDesc, "splitter layer"); - BOOST_TEST(splitterLayer); - - layer->GetOutputSlot(0).Connect(splitterLayer->GetInputSlot(0)); - - // Add a softmax layer 1 - armnn::SoftmaxDescriptor softmaxDescriptor; - armnn::IConnectableLayer* const softmax1Layer = net.AddSoftmaxLayer(softmaxDescriptor, "softmax_1"); - BOOST_TEST(softmax1Layer); - - splitterLayer->GetOutputSlot(0).Connect(softmax1Layer->GetInputSlot(0)); - - // Add a softmax layer 2 - armnn::IConnectableLayer* const softmax2Layer = net.AddSoftmaxLayer(softmaxDescriptor, "softmax_2"); - BOOST_TEST(softmax2Layer); - - splitterLayer->GetOutputSlot(1).Connect(softmax2Layer->GetInputSlot(0)); - - // Add addition layer - layer = net.AddAdditionLayer("add layer"); - BOOST_TEST(layer); - - softmax1Layer->GetOutputSlot(0).Connect(layer->GetInputSlot(0)); - softmax2Layer->GetOutputSlot(0).Connect(layer->GetInputSlot(1)); - - // Add an output layer - armnn::IConnectableLayer* prevLayer = layer; - layer = net.AddOutputLayer(0, "output layer"); - - prevLayer->GetOutputSlot(0).Connect(layer->GetInputSlot(0)); - - BOOST_TEST(layer); -} - -BOOST_AUTO_TEST_CASE(NetworkModification_SplitterMultiplication) -{ - armnn::Network net; - - // Add an input layer and an input tensor descriptor. - armnn::IConnectableLayer* layer = net.AddInputLayer(0, "input layer"); - BOOST_TEST(layer); - - // Add a splitter layer - armnn::ViewsDescriptor splitterDesc(2,4); - armnn::IConnectableLayer* const splitterLayer = net.AddSplitterLayer(splitterDesc, "splitter layer"); - BOOST_TEST(splitterLayer); - - layer->GetOutputSlot(0).Connect(splitterLayer->GetInputSlot(0)); - - // Add a softmax layer 1 - armnn::SoftmaxDescriptor softmaxDescriptor; - armnn::IConnectableLayer* const softmax1Layer = net.AddSoftmaxLayer(softmaxDescriptor, "softmax_1"); - BOOST_TEST(softmax1Layer); - - splitterLayer->GetOutputSlot(0).Connect(softmax1Layer->GetInputSlot(0)); - - // Add a softmax layer 2 - armnn::IConnectableLayer* const softmax2Layer = net.AddSoftmaxLayer(softmaxDescriptor, "softmax_2"); - BOOST_TEST(softmax2Layer); - - splitterLayer->GetOutputSlot(1).Connect(softmax2Layer->GetInputSlot(0)); - - // Add multiplication layer - layer = net.AddMultiplicationLayer("multiplication layer"); - BOOST_TEST(layer); - - softmax1Layer->GetOutputSlot(0).Connect(layer->GetInputSlot(0)); - softmax2Layer->GetOutputSlot(0).Connect(layer->GetInputSlot(1)); - - // Add an output layer - armnn::IConnectableLayer* prevLayer = layer; - layer = net.AddOutputLayer(0, "output layer"); - BOOST_TEST(layer); - - prevLayer->GetOutputSlot(0).Connect(layer->GetInputSlot(0)); -} - -BOOST_AUTO_TEST_CASE(ValidateWorkloads) -{ - const armnn::TensorInfo desc({3, 5}, armnn::DataType::Float32); - - armnn::Network net; - - armnn::NormalizationDescriptor nmDesc; - armnn::ActivationDescriptor acDesc; - - // in - // | - // nm - // / | - // ac | - // \ | - // ml - // | - // sm - // | - // ot - armnn::IConnectableLayer* layer = net.AddInputLayer(0, "in"); - layer->GetOutputSlot(0).SetTensorInfo(desc); - - armnn::IConnectableLayer* const normLayer = net.AddNormalizationLayer(nmDesc, "nm"); - - layer->GetOutputSlot(0).Connect(normLayer->GetInputSlot(0)); - normLayer->GetOutputSlot(0).SetTensorInfo(desc); - - layer = net.AddActivationLayer(acDesc, "ac"); - - normLayer->GetOutputSlot(0).Connect(layer->GetInputSlot(0)); - layer->GetOutputSlot(0).SetTensorInfo(desc); - - armnn::IConnectableLayer* prevLayer = layer; - layer = net.AddMultiplicationLayer("ml"); - - prevLayer->GetOutputSlot(0).Connect(layer->GetInputSlot(0)); - normLayer->GetOutputSlot(0).Connect(layer->GetInputSlot(1)); - layer->GetOutputSlot(0).SetTensorInfo(desc); - - prevLayer = layer; - armnn::SoftmaxDescriptor softmaxDescriptor; - layer = net.AddSoftmaxLayer(softmaxDescriptor, "sm"); - - prevLayer->GetOutputSlot(0).Connect(layer->GetInputSlot(0)); - layer->GetOutputSlot(0).SetTensorInfo(desc); - - prevLayer = layer; - layer = net.AddOutputLayer(0, "ot"); - - prevLayer->GetOutputSlot(0).Connect(layer->GetInputSlot(0)); - - armnn::DeviceSpec spec; - spec.DefaultComputeDevice = armnn::Compute::CpuRef; - - armnn::IOptimizedNetworkPtr optNet = Optimize(net, spec); - static_cast<armnn::OptimizedNetwork*>(optNet.get())->GetGraph().AllocateDynamicBuffers(); - - // validate workloads - armnn::RefWorkloadFactory fact; - for (auto&& layer : static_cast<armnn::OptimizedNetwork*>(optNet.get())->GetGraph()) - { - BOOST_CHECK_NO_THROW( - layer->CreateWorkload(static_cast<armnn::OptimizedNetwork*>(optNet.get())->GetGraph(), fact)); - } -} - -BOOST_AUTO_TEST_SUITE_END() |