// // Copyright © 2017 Arm Ltd. All rights reserved. // SPDX-License-Identifier: MIT // #pragma once #include #include #include #include #include #include #include namespace armnnUtils { template struct ParserPrototxtFixture { ParserPrototxtFixture() : m_Parser(TParser::Create()) , m_NetworkIdentifier(-1) { armnn::IRuntime::CreationOptions options; // Create runtimes for each available backend const armnn::BackendIdSet availableBackendIds = armnn::BackendRegistryInstance().GetBackendIds(); for (auto& backendId : availableBackendIds) { m_Runtimes.push_back(std::make_pair(armnn::IRuntime::Create(options), backendId)); } } /// Parses and loads the network defined by the m_Prototext string. /// @{ void SetupSingleInputSingleOutput(const std::string& inputName, const std::string& outputName); void SetupSingleInputSingleOutput(const armnn::TensorShape& inputTensorShape, const std::string& inputName, const std::string& outputName); void Setup(const std::map& inputShapes, const std::vector& requestedOutputs); void Setup(); /// @} /// Executes the network with the given input tensor and checks the result against the given output tensor. /// This overload assumes that the network has a single input and a single output. template void RunTest(const std::vector& inputData, const std::vector& expectedOutputData); /// Executes the network with the given input tensors and checks the results against the given output tensors. /// This overload supports multiple inputs and multiple outputs, identified by name. template void RunTest(const std::map>& inputData, const std::map>& expectedOutputData); std::string m_Prototext; std::unique_ptr m_Parser; std::vector> m_Runtimes; armnn::NetworkId m_NetworkIdentifier; /// If the single-input-single-output overload of Setup() is called, these will store the input and output name /// so they don't need to be passed to the single-input-single-output overload of RunTest(). /// @{ std::string m_SingleInputName; std::string m_SingleOutputName; /// @} }; template void ParserPrototxtFixture::SetupSingleInputSingleOutput(const std::string& inputName, const std::string& outputName) { // Stores the input and output name so they don't need to be passed to the single-input-single-output RunTest(). m_SingleInputName = inputName; m_SingleOutputName = outputName; Setup({ }, { outputName }); } template void ParserPrototxtFixture::SetupSingleInputSingleOutput(const armnn::TensorShape& inputTensorShape, const std::string& inputName, const std::string& outputName) { // Stores the input and output name so they don't need to be passed to the single-input-single-output RunTest(). m_SingleInputName = inputName; m_SingleOutputName = outputName; Setup({ { inputName, inputTensorShape } }, { outputName }); } template void ParserPrototxtFixture::Setup(const std::map& inputShapes, const std::vector& requestedOutputs) { for (auto&& runtime : m_Runtimes) { std::string errorMessage; armnn::INetworkPtr network = m_Parser->CreateNetworkFromString(m_Prototext.c_str(), inputShapes, requestedOutputs); auto optimized = Optimize(*network, { runtime.second, armnn::Compute::CpuRef }, runtime.first->GetDeviceSpec()); armnn::Status ret = runtime.first->LoadNetwork(m_NetworkIdentifier, move(optimized), errorMessage); if (ret != armnn::Status::Success) { throw armnn::Exception(boost::str( boost::format("LoadNetwork failed with error: '%1%' %2%") % errorMessage % CHECK_LOCATION().AsString())); } } } template void ParserPrototxtFixture::Setup() { for (auto&& runtime : m_Runtimes) { std::string errorMessage; armnn::INetworkPtr network = m_Parser->CreateNetworkFromString(m_Prototext.c_str()); auto optimized = Optimize(*network, { runtime.second, armnn::Compute::CpuRef }, runtime.first->GetDeviceSpec()); armnn::Status ret = runtime.first->LoadNetwork(m_NetworkIdentifier, move(optimized), errorMessage); if (ret != armnn::Status::Success) { throw armnn::Exception(boost::str( boost::format("LoadNetwork failed with error: '%1%' %2%") % errorMessage % CHECK_LOCATION().AsString())); } } } template template void ParserPrototxtFixture::RunTest(const std::vector& inputData, const std::vector& expectedOutputData) { RunTest({ { m_SingleInputName, inputData } }, { { m_SingleOutputName, expectedOutputData } }); } template template void ParserPrototxtFixture::RunTest(const std::map>& inputData, const std::map>& expectedOutputData) { for (auto&& runtime : m_Runtimes) { using BindingPointInfo = std::pair; // Sets up the armnn input tensors from the given vectors. armnn::InputTensors inputTensors; for (auto&& it : inputData) { BindingPointInfo bindingInfo = m_Parser->GetNetworkInputBindingInfo(it.first); inputTensors.push_back({ bindingInfo.first, armnn::ConstTensor(bindingInfo.second, it.second.data()) }); } // Allocates storage for the output tensors to be written to and sets up the armnn output tensors. std::map> outputStorage; armnn::OutputTensors outputTensors; for (auto&& it : expectedOutputData) { BindingPointInfo bindingInfo = m_Parser->GetNetworkOutputBindingInfo(it.first); outputStorage.emplace(it.first, MakeTensor(bindingInfo.second)); outputTensors.push_back( { bindingInfo.first, armnn::Tensor(bindingInfo.second, outputStorage.at(it.first).data()) }); } runtime.first->EnqueueWorkload(m_NetworkIdentifier, inputTensors, outputTensors); // Compares each output tensor to the expected values. for (auto&& it : expectedOutputData) { BindingPointInfo bindingInfo = m_Parser->GetNetworkOutputBindingInfo(it.first); if (bindingInfo.second.GetNumElements() != it.second.size()) { throw armnn::Exception( boost::str( boost::format("Output tensor %1% is expected to have %2% elements. " "%3% elements supplied. %4%") % it.first % bindingInfo.second.GetNumElements() % it.second.size() % CHECK_LOCATION().AsString())); } auto outputExpected = MakeTensor(bindingInfo.second, it.second); BOOST_TEST(CompareTensors(outputExpected, outputStorage[it.first])); } } } } // namespace armnnUtils