diff options
Diffstat (limited to 'src')
-rw-r--r-- | src/armnn/CompatibleTypes.hpp | 2 | ||||
-rw-r--r-- | src/armnn/LayerSupportCommon.hpp | 8 | ||||
-rw-r--r-- | src/armnn/test/CreateWorkload.hpp | 3 | ||||
-rw-r--r-- | src/armnnTfParser/test/Split.cpp | 1 | ||||
-rw-r--r-- | src/armnnUtils/TensorUtils.cpp | 2 | ||||
-rw-r--r-- | src/backends/cl/ClWorkloadFactory.cpp | 7 | ||||
-rw-r--r-- | src/backends/cl/ClWorkloadFactory.hpp | 3 | ||||
-rw-r--r-- | src/backends/neon/NeonWorkloadFactory.cpp | 13 | ||||
-rw-r--r-- | src/backends/neon/NeonWorkloadFactory.hpp | 6 | ||||
-rw-r--r-- | src/backends/reference/RefMemoryManager.cpp | 1 |
10 files changed, 13 insertions, 33 deletions
diff --git a/src/armnn/CompatibleTypes.hpp b/src/armnn/CompatibleTypes.hpp index fd33f6c37a..cc545a9642 100644 --- a/src/armnn/CompatibleTypes.hpp +++ b/src/armnn/CompatibleTypes.hpp @@ -12,7 +12,7 @@ namespace armnn { template<typename T> -bool CompatibleTypes(DataType dataType) +bool CompatibleTypes(DataType) { return false; } diff --git a/src/armnn/LayerSupportCommon.hpp b/src/armnn/LayerSupportCommon.hpp index 8fca3d49d1..557e72a323 100644 --- a/src/armnn/LayerSupportCommon.hpp +++ b/src/armnn/LayerSupportCommon.hpp @@ -70,6 +70,7 @@ bool FalseFunc(Optional<std::string&> reasonIfUnsupported, Params&&... params) template<typename ... Params> bool FalseFuncF16(Optional<std::string&> reasonIfUnsupported, Params&&... params) { + boost::ignore_unused(params...); SetValueChecked(reasonIfUnsupported, "Layer is not supported with float16 data type"); return false; } @@ -77,6 +78,7 @@ bool FalseFuncF16(Optional<std::string&> reasonIfUnsupported, Params&&... params template<typename ... Params> bool FalseFuncF32(Optional<std::string&> reasonIfUnsupported, Params&&... params) { + boost::ignore_unused(params...); SetValueChecked(reasonIfUnsupported, "Layer is not supported with float32 data type"); return false; } @@ -84,6 +86,7 @@ bool FalseFuncF32(Optional<std::string&> reasonIfUnsupported, Params&&... params template<typename ... Params> bool FalseFuncU8(Optional<std::string&> reasonIfUnsupported, Params&&... params) { + boost::ignore_unused(params...); SetValueChecked(reasonIfUnsupported, "Layer is not supported with 8-bit data type"); return false; } @@ -91,6 +94,7 @@ bool FalseFuncU8(Optional<std::string&> reasonIfUnsupported, Params&&... params) template<typename ... Params> bool FalseFuncI32(Optional<std::string&> reasonIfUnsupported, Params&&... params) { + boost::ignore_unused(params...); SetValueChecked(reasonIfUnsupported, "Layer is not supported with int32 data type"); return false; } @@ -98,6 +102,7 @@ bool FalseFuncI32(Optional<std::string&> reasonIfUnsupported, Params&&... params template<typename ... Params> bool FalseInputFuncF32(Optional<std::string&> reasonIfUnsupported, Params&&... params) { + boost::ignore_unused(params...); SetValueChecked(reasonIfUnsupported, "Layer is not supported with float32 data type input"); return false; } @@ -105,6 +110,7 @@ bool FalseInputFuncF32(Optional<std::string&> reasonIfUnsupported, Params&&... p template<typename ... Params> bool FalseInputFuncF16(Optional<std::string&> reasonIfUnsupported, Params&&... params) { + boost::ignore_unused(params...); SetValueChecked(reasonIfUnsupported, "Layer is not supported with float16 data type input"); return false; } @@ -112,6 +118,7 @@ bool FalseInputFuncF16(Optional<std::string&> reasonIfUnsupported, Params&&... p template<typename ... Params> bool FalseOutputFuncF32(Optional<std::string&> reasonIfUnsupported, Params&&... params) { + boost::ignore_unused(params...); SetValueChecked(reasonIfUnsupported, "Layer is not supported with float32 data type output"); return false; } @@ -119,6 +126,7 @@ bool FalseOutputFuncF32(Optional<std::string&> reasonIfUnsupported, Params&&... template<typename ... Params> bool FalseOutputFuncF16(Optional<std::string&> reasonIfUnsupported, Params&&... params) { + boost::ignore_unused(params...); SetValueChecked(reasonIfUnsupported, "Layer is not supported with float16 data type output"); return false; } diff --git a/src/armnn/test/CreateWorkload.hpp b/src/armnn/test/CreateWorkload.hpp index 4e7967bf40..f6928f858f 100644 --- a/src/armnn/test/CreateWorkload.hpp +++ b/src/armnn/test/CreateWorkload.hpp @@ -18,6 +18,7 @@ #include <boost/test/unit_test.hpp> #include <boost/cast.hpp> +#include <boost/core/ignore_unused.hpp> #include <utility> @@ -1324,6 +1325,8 @@ std::pair<armnn::IOptimizedNetworkPtr, std::unique_ptr<PreCompiledWorkload>> Cre armnn::Graph& graph, bool biasEnabled = false) { + boost::ignore_unused(graph); + // To create a PreCompiled layer, create a network and Optimize it. armnn::Network net; diff --git a/src/armnnTfParser/test/Split.cpp b/src/armnnTfParser/test/Split.cpp index 10ff04df89..d53ae672eb 100644 --- a/src/armnnTfParser/test/Split.cpp +++ b/src/armnnTfParser/test/Split.cpp @@ -176,6 +176,7 @@ BOOST_FIXTURE_TEST_CASE(ParseSplit, InputFirstSplitFixture) struct SplitLastDimFixture : public armnnUtils::ParserPrototxtFixture<armnnTfParser::ITfParser> { SplitLastDimFixture(bool withDimZero=false) { + boost::ignore_unused(withDimZero); m_Prototext = R"( node { name: "Placeholder" diff --git a/src/armnnUtils/TensorUtils.cpp b/src/armnnUtils/TensorUtils.cpp index 84fc8db053..535d68adbe 100644 --- a/src/armnnUtils/TensorUtils.cpp +++ b/src/armnnUtils/TensorUtils.cpp @@ -114,7 +114,6 @@ unsigned int GetNumElementsBetween(const TensorShape& shape, const unsigned int firstAxisInclusive, const unsigned int lastAxisExclusive) { - BOOST_ASSERT(0 <= firstAxisInclusive); BOOST_ASSERT(firstAxisInclusive <= lastAxisExclusive); BOOST_ASSERT(lastAxisExclusive <= shape.GetNumDimensions()); unsigned int count = 1; @@ -141,7 +140,6 @@ unsigned int GetUnsignedAxis(const unsigned int inputDimension, const int axis) unsigned int GetNumElementsAfter(const armnn::TensorShape& shape, unsigned int axis) { unsigned int numDim = shape.GetNumDimensions(); - BOOST_ASSERT(0 >= axis); BOOST_ASSERT(axis <= numDim - 1); unsigned int count = 1; for (unsigned int i = axis; i < numDim; i++) diff --git a/src/backends/cl/ClWorkloadFactory.cpp b/src/backends/cl/ClWorkloadFactory.cpp index 4746167795..f9e6632b0c 100644 --- a/src/backends/cl/ClWorkloadFactory.cpp +++ b/src/backends/cl/ClWorkloadFactory.cpp @@ -260,13 +260,6 @@ std::unique_ptr<IWorkload> ClWorkloadFactory::CreateEqual(const EqualQueueDescri return CreateComparison(comparisonDescriptor, info); } -std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFakeQuantization( - const FakeQuantizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const -{ - return MakeWorkload<NullWorkload, NullWorkload>(descriptor, info); -} - std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info) const { diff --git a/src/backends/cl/ClWorkloadFactory.hpp b/src/backends/cl/ClWorkloadFactory.hpp index 8c94818db2..8f377e959d 100644 --- a/src/backends/cl/ClWorkloadFactory.hpp +++ b/src/backends/cl/ClWorkloadFactory.hpp @@ -95,9 +95,6 @@ public: std::unique_ptr<IWorkload> CreateEqual(const EqualQueueDescriptor& descriptor, const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateFloor(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info) const override; diff --git a/src/backends/neon/NeonWorkloadFactory.cpp b/src/backends/neon/NeonWorkloadFactory.cpp index 8d798ec864..1cc9e50e0b 100644 --- a/src/backends/neon/NeonWorkloadFactory.cpp +++ b/src/backends/neon/NeonWorkloadFactory.cpp @@ -225,13 +225,6 @@ std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateEqual(const EqualQueueDesc return CreateComparison(comparisonDescriptor, info); } -std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateFakeQuantization( - const FakeQuantizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const -{ - return nullptr; -} - std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info) const { @@ -441,12 +434,6 @@ std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateSoftmax(const SoftmaxQueue descriptor, info, m_MemoryManager->GetIntraLayerManager()); } -std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor, - const WorkloadInfo& info) const -{ - return nullptr; -} - std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateSpaceToDepth( const armnn::SpaceToDepthQueueDescriptor& descriptor, const armnn::WorkloadInfo& info) const { diff --git a/src/backends/neon/NeonWorkloadFactory.hpp b/src/backends/neon/NeonWorkloadFactory.hpp index 6bdc237fff..b76a3a340a 100644 --- a/src/backends/neon/NeonWorkloadFactory.hpp +++ b/src/backends/neon/NeonWorkloadFactory.hpp @@ -97,9 +97,6 @@ public: std::unique_ptr<IWorkload> CreateEqual(const EqualQueueDescriptor& descriptor, const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateFloor(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info) const override; @@ -193,9 +190,6 @@ public: std::unique_ptr<IWorkload> CreateSoftmax(const SoftmaxQueueDescriptor& descriptor, const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor, - const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& descriptor, const WorkloadInfo& info) const override; diff --git a/src/backends/reference/RefMemoryManager.cpp b/src/backends/reference/RefMemoryManager.cpp index fdd008dbe6..4f15e39ee1 100644 --- a/src/backends/reference/RefMemoryManager.cpp +++ b/src/backends/reference/RefMemoryManager.cpp @@ -88,7 +88,6 @@ void RefMemoryManager::Pool::Reserve(unsigned int numBytes) void RefMemoryManager::Pool::Acquire() { BOOST_ASSERT_MSG(!m_Pointer, "RefMemoryManager::Pool::Acquire() called when memory already acquired"); - BOOST_ASSERT(m_Size >= 0); m_Pointer = ::operator new(size_t(m_Size)); } |