16 #include <boost/test/unit_test.hpp> 17 #include <boost/algorithm/string/trim.hpp> 24 bool LayerTypeMatchesTest()
26 return LayerTypeMatchesTestImpl<armnn::LayerType::FirstLayer>(Tag<armnn::LayerType::FirstLayer>());
35 LayerTypeMatchesTest();
47 std::string reasonNotSupported;
54 IsLayerSupportedTests<armnn::RefWorkloadFactory, armnn::DataType::Float16>(&factory);
60 IsLayerSupportedTests<armnn::RefWorkloadFactory, armnn::DataType::Float32>(&factory);
66 IsLayerSupportedTests<armnn::RefWorkloadFactory, armnn::DataType::QAsymmU8>(&factory);
72 IsLayerSupportedTests<armnn::RefWorkloadFactory, armnn::DataType::QSymmS8>(&factory);
78 IsLayerSupportedTests<armnn::RefWorkloadFactory, armnn::DataType::QSymmS16>(&factory);
83 std::string reasonIfUnsupported;
93 std::string reasonIfUnsupported;
99 BOOST_CHECK_EQUAL(reasonIfUnsupported,
"Layer is not supported with float32 data type input");
104 std::string reasonIfUnsupported;
110 BOOST_CHECK_EQUAL(reasonIfUnsupported,
"Layer is not supported with float16 data type output");
115 std::string reasonIfUnsupported;
125 std::string reasonIfUnsupported;
131 BOOST_CHECK_EQUAL(reasonIfUnsupported,
"Layer is not supported with float16 data type input");
136 std::string reasonIfUnsupported;
142 BOOST_CHECK_EQUAL(reasonIfUnsupported,
"Layer is not supported with float32 data type output");
147 std::string reasonIfUnsupported;
157 std::string reasonIfUnsupported;
164 boost::algorithm::trim(reasonIfUnsupported);
165 BOOST_CHECK_EQUAL(reasonIfUnsupported,
166 "Reference Mean: Expected 4 dimensions but got 2 dimensions instead, for the 'output' tensor.");
bool IsAdditionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
This layer converts data type Float 32 to Float 16.
BOOST_CHECK(profilingService.GetCurrentState()==ProfilingState::WaitingForAck)
BOOST_AUTO_TEST_SUITE_END()
BOOST_AUTO_TEST_CASE(IsLayerSupportedLayerTypeMatches)
BOOST_AUTO_TEST_SUITE(TensorflowLiteParser)
This layer converts data type Float 16 to Float 32.