From 901ea11e3da6b97df52bdc2a547990402e920ede Mon Sep 17 00:00:00 2001 From: Derek Lamberti Date: Tue, 10 Dec 2019 22:07:09 +0000 Subject: IVGCVSW-4246 Clean build of backends with -Wextra Change-Id: I9e8d5576b3ec04c871785d5f2f9545bf1136e59b Signed-off-by: Derek Lamberti --- include/armnn/backends/IBackendInternal.hpp | 2 +- include/armnn/backends/ITensorHandle.hpp | 8 +- src/backends/backendsCommon/IBackendInternal.cpp | 6 +- src/backends/backendsCommon/LayerSupportBase.cpp | 399 ++++++++++----------- src/backends/backendsCommon/LayerSupportRules.hpp | 2 +- src/backends/backendsCommon/MakeWorkloadHelper.hpp | 3 + src/backends/backendsCommon/Workload.hpp | 2 +- src/backends/backendsCommon/WorkloadData.cpp | 2 +- src/backends/backendsCommon/WorkloadFactory.cpp | 230 ++++++------ .../backendsCommon/WorkloadFactoryBase.hpp | 238 ++++++------ .../test/IsLayerSupportedTestImpl.hpp | 1 + src/backends/backendsCommon/test/MockBackend.cpp | 2 +- src/backends/backendsCommon/test/MockBackend.hpp | 28 +- .../backendsCommon/test/TestDynamicBackend.cpp | 3 + src/backends/cl/ClLayerSupport.cpp | 11 +- src/backends/cl/ClWorkloadFactory.cpp | 3 + src/backends/cl/OpenClTimer.cpp | 3 + src/backends/neon/NeonLayerSupport.cpp | 11 +- src/backends/reference/RefLayerSupport.cpp | 18 +- src/backends/reference/RefWorkloadFactory.cpp | 12 +- .../reference/test/RefWorkloadFactoryHelper.hpp | 1 + src/backends/reference/workloads/ArgMinMax.cpp | 2 + src/backends/reference/workloads/Dequantize.cpp | 2 + .../reference/workloads/DetectionPostProcess.cpp | 2 + src/backends/reference/workloads/Gather.cpp | 2 + src/backends/reference/workloads/Pooling2d.cpp | 2 +- 26 files changed, 518 insertions(+), 477 deletions(-) diff --git a/include/armnn/backends/IBackendInternal.hpp b/include/armnn/backends/IBackendInternal.hpp index 3296d81b7c..3533aceb3c 100644 --- a/include/armnn/backends/IBackendInternal.hpp +++ b/include/armnn/backends/IBackendInternal.hpp @@ -127,7 +127,7 @@ public: /// (Optional) Register TensorHandleFactories /// Either this method or CreateMemoryManager() and /// IWorkloadFactory::CreateTensor()/IWorkloadFactory::CreateSubtensor() methods must be implemented. - virtual void RegisterTensorHandleFactories(class TensorHandleFactoryRegistry& registry) {} + virtual void RegisterTensorHandleFactories(class TensorHandleFactoryRegistry& /*registry*/) {} /// Returns the version of the Backend API static constexpr BackendVersion GetApiVersion() { return BackendVersion(1, 0); } diff --git a/include/armnn/backends/ITensorHandle.hpp b/include/armnn/backends/ITensorHandle.hpp index e1b80b874a..6ef0e32fb3 100644 --- a/include/armnn/backends/ITensorHandle.hpp +++ b/include/armnn/backends/ITensorHandle.hpp @@ -6,6 +6,8 @@ #include +#include + namespace armnn { @@ -71,7 +73,11 @@ public: /// \param memory base address of the memory being imported. /// \param source source of the allocation for the memory being imported. /// \return true on success or false on failure - virtual bool Import(void* memory, MemorySource source) { return false; }; + virtual bool Import(void* memory, MemorySource source) + { + boost::ignore_unused(memory, source); + return false; + }; }; } diff --git a/src/backends/backendsCommon/IBackendInternal.cpp b/src/backends/backendsCommon/IBackendInternal.cpp index d1422d1c22..ad09730e33 100644 --- a/src/backends/backendsCommon/IBackendInternal.cpp +++ b/src/backends/backendsCommon/IBackendInternal.cpp @@ -10,7 +10,7 @@ namespace armnn ARMNN_NO_DEPRECATE_WARN_BEGIN IBackendInternal::ISubGraphConverterPtr IBackendInternal::CreateSubGraphConverter( - const std::shared_ptr& subGraph) const + const std::shared_ptr& /*subGrapg*/) const { return ISubGraphConverterPtr{}; } @@ -20,7 +20,7 @@ IBackendInternal::Optimizations IBackendInternal::GetOptimizations() const return Optimizations{}; } -IBackendInternal::SubGraphUniquePtr IBackendInternal::OptimizeSubGraph(const SubGraph& subGraph, +IBackendInternal::SubGraphUniquePtr IBackendInternal::OptimizeSubGraph(const SubGraph& /*subGraph*/, bool& optimizationAttempted) const { optimizationAttempted = false; @@ -34,7 +34,7 @@ IMemoryManagerUniquePtr IBackendInternal::CreateMemoryManager() const } IBackendInternal::IWorkloadFactoryPtr IBackendInternal::CreateWorkloadFactory( - class TensorHandleFactoryRegistry& tensorHandleFactoryRegistry) const + class TensorHandleFactoryRegistry& /*tensorHandleFactoryRegistry*/) const { return IWorkloadFactoryPtr{}; } diff --git a/src/backends/backendsCommon/LayerSupportBase.cpp b/src/backends/backendsCommon/LayerSupportBase.cpp index 00f1d0223d..8332774202 100644 --- a/src/backends/backendsCommon/LayerSupportBase.cpp +++ b/src/backends/backendsCommon/LayerSupportBase.cpp @@ -34,295 +34,292 @@ bool DefaultLayerSupport(const char* func, namespace armnn { -bool LayerSupportBase::IsAbsSupported(const TensorInfo &input, - const TensorInfo &output, +bool LayerSupportBase::IsAbsSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsActivationSupported(const TensorInfo& input, - const TensorInfo& output, - const ActivationDescriptor& descriptor, +bool LayerSupportBase::IsActivationSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const ActivationDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsAdditionSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, +bool LayerSupportBase::IsAdditionSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsArgMinMaxSupported(const armnn::TensorInfo &input, const armnn::TensorInfo &output, - const armnn::ArgMinMaxDescriptor& descriptor, +bool LayerSupportBase::IsArgMinMaxSupported(const armnn::TensorInfo &/*input*/, + const armnn::TensorInfo &/*output*/, + const armnn::ArgMinMaxDescriptor& /*descriptor*/, armnn::Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsBatchNormalizationSupported(const TensorInfo& input, - const TensorInfo& output, - const TensorInfo& mean, - const TensorInfo& var, - const TensorInfo& beta, - const TensorInfo& gamma, - const BatchNormalizationDescriptor& descriptor, +bool LayerSupportBase::IsBatchNormalizationSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const TensorInfo& /*mean*/, + const TensorInfo& /*var*/, + const TensorInfo& /*beta*/, + const TensorInfo& /*gamma*/, + const BatchNormalizationDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsBatchToSpaceNdSupported(const TensorInfo& input, - const TensorInfo& output, - const BatchToSpaceNdDescriptor& descriptor, +bool LayerSupportBase::IsBatchToSpaceNdSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const BatchToSpaceNdDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsComparisonSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, - const ComparisonDescriptor& descriptor, +bool LayerSupportBase::IsComparisonSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, + const ComparisonDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsConcatSupported(const std::vector inputs, - const TensorInfo& output, - const OriginsDescriptor& descriptor, +bool LayerSupportBase::IsConcatSupported(const std::vector /*inputs*/, + const TensorInfo& /*output*/, + const OriginsDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsConstantSupported(const TensorInfo& output, +bool LayerSupportBase::IsConstantSupported(const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsConvertFp16ToFp32Supported(const TensorInfo& input, - const TensorInfo& output, +bool LayerSupportBase::IsConvertFp16ToFp32Supported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsConvertFp32ToFp16Supported(const TensorInfo& input, - const TensorInfo& output, +bool LayerSupportBase::IsConvertFp32ToFp16Supported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsConvolution2dSupported(const TensorInfo& input, - const TensorInfo& output, - const Convolution2dDescriptor& descriptor, - const TensorInfo& weights, - const Optional& biases, +bool LayerSupportBase::IsConvolution2dSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const Convolution2dDescriptor& /*descriptor*/, + const TensorInfo& /*weights*/, + const Optional& /*biases*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsDebugSupported(const TensorInfo& input, - const TensorInfo& output, +bool LayerSupportBase::IsDebugSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsDepthToSpaceSupported(const TensorInfo& input, - const TensorInfo& output, - const DepthToSpaceDescriptor& descriptor, +bool LayerSupportBase::IsDepthToSpaceSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const DepthToSpaceDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsDepthwiseConvolutionSupported(const TensorInfo& input, - const TensorInfo& output, - const DepthwiseConvolution2dDescriptor& descriptor, - const TensorInfo& weights, - const Optional& biases, +bool LayerSupportBase::IsDepthwiseConvolutionSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const DepthwiseConvolution2dDescriptor& /*descriptor*/, + const TensorInfo& /*weights*/, + const Optional& /*biases*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsDequantizeSupported(const TensorInfo& input, - const TensorInfo& output, +bool LayerSupportBase::IsDequantizeSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsDetectionPostProcessSupported(const TensorInfo& boxEncodings, - const TensorInfo& scores, - const TensorInfo& anchors, - const TensorInfo& detectionBoxes, - const TensorInfo& detectionClasses, - const TensorInfo& detectionScores, - const TensorInfo& numDetections, - const DetectionPostProcessDescriptor& descriptor, +bool LayerSupportBase::IsDetectionPostProcessSupported(const TensorInfo& /*boxEncodings*/, + const TensorInfo& /*scores*/, + const TensorInfo& /*anchors*/, + const TensorInfo& /*detectionBoxes*/, + const TensorInfo& /*detectionClasses*/, + const TensorInfo& /*detectionScores*/, + const TensorInfo& /*numDetections*/, + const DetectionPostProcessDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsDilatedDepthwiseConvolutionSupported(const TensorInfo& input, - const TensorInfo& output, - const DepthwiseConvolution2dDescriptor& descriptor, - const TensorInfo& weights, - const Optional& biases, +bool LayerSupportBase::IsDilatedDepthwiseConvolutionSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const DepthwiseConvolution2dDescriptor& /*descriptor*/, + const TensorInfo& /*weights*/, + const Optional& /*biases*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsDivisionSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, +bool LayerSupportBase::IsDivisionSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsEqualSupported(const armnn::TensorInfo& input0, - const armnn::TensorInfo& input1, - const armnn::TensorInfo& output, +bool LayerSupportBase::IsEqualSupported(const armnn::TensorInfo& /*input0*/, + const armnn::TensorInfo& /*input1*/, + const armnn::TensorInfo& /*output*/, armnn::Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsFakeQuantizationSupported(const TensorInfo& input, - const FakeQuantizationDescriptor& descriptor, +bool LayerSupportBase::IsFakeQuantizationSupported(const TensorInfo& /*input*/, + const FakeQuantizationDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsFloorSupported(const TensorInfo& input, - const TensorInfo& output, +bool LayerSupportBase::IsFloorSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsFullyConnectedSupported(const TensorInfo& input, - const TensorInfo& output, - const TensorInfo& weights, - const TensorInfo& biases, - const FullyConnectedDescriptor& descriptor, +bool LayerSupportBase::IsFullyConnectedSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const TensorInfo& /*weights*/, + const TensorInfo& /*biases*/, + const FullyConnectedDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsGatherSupported(const armnn::TensorInfo& input0, - const armnn::TensorInfo& input1, - const armnn::TensorInfo& output, +bool LayerSupportBase::IsGatherSupported(const armnn::TensorInfo& /*input0*/, + const armnn::TensorInfo& /*input1*/, + const armnn::TensorInfo& /*output*/, armnn::Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsGreaterSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, +bool LayerSupportBase::IsGreaterSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsInputSupported(const TensorInfo& input, +bool LayerSupportBase::IsInputSupported(const TensorInfo& /*input*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsInstanceNormalizationSupported(const TensorInfo& input, - const TensorInfo& output, - const InstanceNormalizationDescriptor& descriptor, +bool LayerSupportBase::IsInstanceNormalizationSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const InstanceNormalizationDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsL2NormalizationSupported(const TensorInfo& input, - const TensorInfo& output, - const L2NormalizationDescriptor& descriptor, +bool LayerSupportBase::IsL2NormalizationSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const L2NormalizationDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsLogSoftmaxSupported(const TensorInfo& input, - const TensorInfo& output, - const LogSoftmaxDescriptor& descriptor, +bool LayerSupportBase::IsLogSoftmaxSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const LogSoftmaxDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsLstmSupported(const TensorInfo& input, - const TensorInfo& outputStateIn, - const TensorInfo& cellStateIn, - const TensorInfo& scratchBuffer, - const TensorInfo& outputStateOut, - const TensorInfo& cellStateOut, - const TensorInfo& output, - const LstmDescriptor& descriptor, - const LstmInputParamsInfo& paramsInfo, +bool LayerSupportBase::IsLstmSupported(const TensorInfo& /*input*/, + const TensorInfo& /*outputStateIn*/, + const TensorInfo& /*cellStateIn*/, + const TensorInfo& /*scratchBuffer*/, + const TensorInfo& /*outputStateOut*/, + const TensorInfo& /*cellStateOut*/, + const TensorInfo& /*output*/, + const LstmDescriptor& /*descriptor*/, + const LstmInputParamsInfo& /*paramsInfo*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsMaximumSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, +bool LayerSupportBase::IsMaximumSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsMeanSupported(const TensorInfo& input, - const TensorInfo& output, - const MeanDescriptor& descriptor, +bool LayerSupportBase::IsMeanSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const MeanDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsMemCopySupported(const armnn::TensorInfo& input, - const armnn::TensorInfo& output, - armnn::Optional reasonIfUnsupported) const +bool LayerSupportBase::IsMemCopySupported(const armnn::TensorInfo& /*input*/, + const armnn::TensorInfo& /*output*/, + armnn::Optional /*reasonIfUnsupported*/) const { - boost::ignore_unused(input); - boost::ignore_unused(output); return true; } -bool LayerSupportBase::IsMemImportSupported(const armnn::TensorInfo& input, - const armnn::TensorInfo& output, - armnn::Optional reasonIfUnsupported) const +bool LayerSupportBase::IsMemImportSupported(const armnn::TensorInfo& /*input*/, + const armnn::TensorInfo& /*output*/, + armnn::Optional /*reasonIfUnsupported*/) const { - boost::ignore_unused(input); - boost::ignore_unused(output); return true; } -bool LayerSupportBase::IsMergeSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, +bool LayerSupportBase::IsMergeSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); @@ -336,180 +333,180 @@ bool LayerSupportBase::IsMergerSupported(const std::vector in return IsConcatSupported(inputs, output, descriptor, reasonIfUnsupported); } -bool LayerSupportBase::IsMinimumSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, +bool LayerSupportBase::IsMinimumSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsMultiplicationSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, +bool LayerSupportBase::IsMultiplicationSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsNormalizationSupported(const TensorInfo& input, - const TensorInfo& output, - const NormalizationDescriptor& descriptor, +bool LayerSupportBase::IsNormalizationSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const NormalizationDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsOutputSupported(const TensorInfo& output, +bool LayerSupportBase::IsOutputSupported(const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsPadSupported(const TensorInfo& input, - const TensorInfo& output, - const PadDescriptor& descriptor, +bool LayerSupportBase::IsPadSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const PadDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsPermuteSupported(const TensorInfo& input, - const TensorInfo& output, - const PermuteDescriptor& descriptor, +bool LayerSupportBase::IsPermuteSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const PermuteDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsPooling2dSupported(const TensorInfo& input, - const TensorInfo& output, - const Pooling2dDescriptor& descriptor, +bool LayerSupportBase::IsPooling2dSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const Pooling2dDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsPreCompiledSupported(const TensorInfo& input, - const PreCompiledDescriptor& descriptor, +bool LayerSupportBase::IsPreCompiledSupported(const TensorInfo& /*input*/, + const PreCompiledDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsPreluSupported(const TensorInfo& input, - const TensorInfo& alpha, - const TensorInfo& output, +bool LayerSupportBase::IsPreluSupported(const TensorInfo& /*input*/, + const TensorInfo& /*alpha*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsQuantizeSupported(const armnn::TensorInfo& input, - const armnn::TensorInfo& output, +bool LayerSupportBase::IsQuantizeSupported(const armnn::TensorInfo& /*input*/, + const armnn::TensorInfo& /*output*/, armnn::Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsQuantizedLstmSupported(const TensorInfo& input, - const TensorInfo& previousCellStateIn, - const TensorInfo& previousOutputIn, - const TensorInfo& cellStateOut, - const TensorInfo& output, - const QuantizedLstmInputParamsInfo& paramsInfo, +bool LayerSupportBase::IsQuantizedLstmSupported(const TensorInfo& /*input*/, + const TensorInfo& /*previousCellStateIn*/, + const TensorInfo& /*previousOutputIn*/, + const TensorInfo& /*cellStateOut*/, + const TensorInfo& /*output*/, + const QuantizedLstmInputParamsInfo& /*paramsInfo*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsReshapeSupported(const TensorInfo& input, - const ReshapeDescriptor& descriptor, +bool LayerSupportBase::IsReshapeSupported(const TensorInfo& /*input*/, + const ReshapeDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsResizeBilinearSupported(const TensorInfo& input, - const TensorInfo& output, +bool LayerSupportBase::IsResizeBilinearSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsResizeSupported(const TensorInfo& input, - const TensorInfo& output, - const ResizeDescriptor& descriptor, +bool LayerSupportBase::IsResizeSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const ResizeDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsRsqrtSupported(const TensorInfo &input, - const TensorInfo &output, +bool LayerSupportBase::IsRsqrtSupported(const TensorInfo &/*input*/, + const TensorInfo &/*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsSliceSupported(const TensorInfo& input, - const TensorInfo& output, - const SliceDescriptor& descriptor, +bool LayerSupportBase::IsSliceSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const SliceDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsSoftmaxSupported(const TensorInfo& input, - const TensorInfo& output, - const SoftmaxDescriptor& descriptor, +bool LayerSupportBase::IsSoftmaxSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const SoftmaxDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } - -bool LayerSupportBase::IsSpaceToBatchNdSupported(const TensorInfo& input, - const TensorInfo& output, - const SpaceToBatchNdDescriptor& descriptor, +/**/ +bool LayerSupportBase::IsSpaceToBatchNdSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const SpaceToBatchNdDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsSpaceToDepthSupported(const TensorInfo& input, - const TensorInfo& output, - const SpaceToDepthDescriptor& descriptor, +bool LayerSupportBase::IsSpaceToDepthSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const SpaceToDepthDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsSplitterSupported(const TensorInfo& input, - const ViewsDescriptor& descriptor, +bool LayerSupportBase::IsSplitterSupported(const TensorInfo& /*input*/, + const ViewsDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsSplitterSupported(const TensorInfo& input, - const std::vector>& outputs, - const ViewsDescriptor& descriptor, +bool LayerSupportBase::IsSplitterSupported(const TensorInfo& /*input*/, + const std::vector>& /*outputs*/, + const ViewsDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsStackSupported(const std::vector& inputs, - const TensorInfo& output, - const StackDescriptor& descriptor, +bool LayerSupportBase::IsStackSupported(const std::vector& /*inputs*/, + const TensorInfo& /*output*/, + const StackDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsStandInSupported(const std::vector& inputs, - const std::vector& outputs, - const StandInDescriptor& descriptor, +bool LayerSupportBase::IsStandInSupported(const std::vector& /*inputs*/, + const std::vector& /*outputs*/, + const StandInDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { if (reasonIfUnsupported) @@ -523,36 +520,36 @@ bool LayerSupportBase::IsStandInSupported(const std::vector& return false; } -bool LayerSupportBase::IsStridedSliceSupported(const TensorInfo& input, - const TensorInfo& output, - const StridedSliceDescriptor& descriptor, +bool LayerSupportBase::IsStridedSliceSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const StridedSliceDescriptor& /*descriptor*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsSubtractionSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, +bool LayerSupportBase::IsSubtractionSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsSwitchSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output0, - const TensorInfo& output1, +bool LayerSupportBase::IsSwitchSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output0*/, + const TensorInfo& /*output1*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsTransposeConvolution2dSupported(const TensorInfo& input, - const TensorInfo& output, - const TransposeConvolution2dDescriptor& descriptor, - const TensorInfo& weights, - const Optional& biases, +bool LayerSupportBase::IsTransposeConvolution2dSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const TransposeConvolution2dDescriptor& /*descriptor*/, + const TensorInfo& /*weights*/, + const Optional& /*biases*/, Optional reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); diff --git a/src/backends/backendsCommon/LayerSupportRules.hpp b/src/backends/backendsCommon/LayerSupportRules.hpp index bf997dbff7..08189f9999 100644 --- a/src/backends/backendsCommon/LayerSupportRules.hpp +++ b/src/backends/backendsCommon/LayerSupportRules.hpp @@ -55,7 +55,7 @@ struct Rule }; template -bool AllTypesAreEqualImpl(T t) +bool AllTypesAreEqualImpl(T) { return true; } diff --git a/src/backends/backendsCommon/MakeWorkloadHelper.hpp b/src/backends/backendsCommon/MakeWorkloadHelper.hpp index 71358bbbef..9d8174ce7d 100644 --- a/src/backends/backendsCommon/MakeWorkloadHelper.hpp +++ b/src/backends/backendsCommon/MakeWorkloadHelper.hpp @@ -31,6 +31,9 @@ struct MakeWorkloadForType const WorkloadInfo& info, Args&&... args) { + boost::ignore_unused(descriptor); + boost::ignore_unused(info); + boost::ignore_unused(args...); return nullptr; } }; diff --git a/src/backends/backendsCommon/Workload.hpp b/src/backends/backendsCommon/Workload.hpp index b5851ad7b9..e03068618b 100644 --- a/src/backends/backendsCommon/Workload.hpp +++ b/src/backends/backendsCommon/Workload.hpp @@ -27,7 +27,7 @@ public: virtual profiling::ProfilingGuid GetGuid() const = 0; - virtual void RegisterDebugCallback(const DebugCallbackFunction& func) {} + virtual void RegisterDebugCallback(const DebugCallbackFunction& /*func*/) {} }; // NullWorkload used to denote an unsupported workload when used by the MakeWorkload<> template diff --git a/src/backends/backendsCommon/WorkloadData.cpp b/src/backends/backendsCommon/WorkloadData.cpp index d9a1f46c9f..c3dd601fbd 100644 --- a/src/backends/backendsCommon/WorkloadData.cpp +++ b/src/backends/backendsCommon/WorkloadData.cpp @@ -2554,7 +2554,7 @@ void SwitchQueueDescriptor::Validate(const WorkloadInfo& workloadInfo) const "output_1"); } -void PreCompiledQueueDescriptor::Validate(const WorkloadInfo& workloadInfo) const +void PreCompiledQueueDescriptor::Validate(const WorkloadInfo& /*workloadInfo*/) const { // This is internally generated so it should not need validation. } diff --git a/src/backends/backendsCommon/WorkloadFactory.cpp b/src/backends/backendsCommon/WorkloadFactory.cpp index 805ec7ba5f..a4327e441a 100644 --- a/src/backends/backendsCommon/WorkloadFactory.cpp +++ b/src/backends/backendsCommon/WorkloadFactory.cpp @@ -1074,358 +1074,358 @@ bool IWorkloadFactory::IsLayerSupported(const IConnectableLayer& connectableLaye } // Default Implementations -std::unique_ptr IWorkloadFactory::CreateAbs(const AbsQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateAbs(const AbsQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateArgMinMax(const ArgMinMaxQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateArgMinMax(const ArgMinMaxQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } std::unique_ptr IWorkloadFactory::CreateBatchNormalization( - const BatchNormalizationQueueDescriptor& descriptor, const WorkloadInfo& info) const + const BatchNormalizationQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& descriptor, - const WorkloadInfo& Info) const +std::unique_ptr IWorkloadFactory::CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& /*desc*/, + const WorkloadInfo& /*Info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateComparison(const ComparisonQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateComparison(const ComparisonQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateConcat(const ConcatQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateConcat(const ConcatQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateConvertFp16ToFp32(const ConvertFp16ToFp32QueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateConvertFp16ToFp32(const ConvertFp16ToFp32QueueDescriptor& /*desc*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateConvertFp32ToFp16(const ConvertFp32ToFp16QueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateConvertFp32ToFp16(const ConvertFp32ToFp16QueueDescriptor& /*desc*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateDebug(const DebugQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateDebug(const DebugQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateDepthToSpace(const DepthToSpaceQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateDepthToSpace(const DepthToSpaceQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } std::unique_ptr IWorkloadFactory::CreateDepthwiseConvolution2d( - const DepthwiseConvolution2dQueueDescriptor& descriptor, const WorkloadInfo& info) const + const DepthwiseConvolution2dQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } std::unique_ptr IWorkloadFactory::CreateDequantize( - const DequantizeQueueDescriptor& descriptor, const WorkloadInfo& info) const + const DequantizeQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } std::unique_ptr IWorkloadFactory::CreateDetectionPostProcess( - const DetectionPostProcessQueueDescriptor& descriptor, const WorkloadInfo& info) const + const DetectionPostProcessQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateDivision(const DivisionQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateDivision(const DivisionQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor, - const WorkloadInfo& Info) const +std::unique_ptr IWorkloadFactory::CreateEqual(const EqualQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*Info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateFakeQuantization(const FakeQuantizationQueueDescriptor& /*desc*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateFloor(const FloorQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateFullyConnected(const FullyConnectedQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateFullyConnected(const FullyConnectedQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateGather(const GatherQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateGather(const GatherQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } std::unique_ptr IWorkloadFactory::CreateInstanceNormalization( - const InstanceNormalizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const + const InstanceNormalizationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& /*desc*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateLogSoftmax(const LogSoftmaxQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateLogSoftmax(const LogSoftmaxQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateLstm(const LstmQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateMaximum(const MaximumQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateMaximum(const MaximumQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateMean(const MeanQueueDescriptor& descriptor, - const WorkloadInfo& Info) const +std::unique_ptr IWorkloadFactory::CreateMean(const MeanQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*Info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateMemImport(const MemImportQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateMemImport(const MemImportQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateMerge(const MergeQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateMerge(const MergeQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateMerger(const MergerQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateMerger(const MergerQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateMultiplication(const MultiplicationQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateMultiplication(const MultiplicationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateOutput(const OutputQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateOutput(const OutputQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreatePad(const PadQueueDescriptor& descriptor, - const WorkloadInfo& Info) const +std::unique_ptr IWorkloadFactory::CreatePad(const PadQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*Info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& /*descriptor*/, + const WorkloadInfo&/**/ /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreatePrelu(const PreluQueueDescriptor &descriptor, - const WorkloadInfo &info) const +std::unique_ptr IWorkloadFactory::CreatePrelu(const PreluQueueDescriptor &/*descriptor*/, + const WorkloadInfo &/*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateQuantize(const QuantizeQueueDescriptor& descriptor, - const WorkloadInfo& Info) const +std::unique_ptr IWorkloadFactory::CreateQuantize(const QuantizeQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*Info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateQuantizedLstm(const QuantizedLstmQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateQuantizedLstm(const QuantizedLstmQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateResizeBilinear(const ResizeBilinearQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateResizeBilinear(const ResizeBilinearQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateResize(const ResizeQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateResize(const ResizeQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateSlice(const SliceQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateSlice(const SliceQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } - -std::unique_ptr IWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor, - const WorkloadInfo& info) const +/**/ +std::unique_ptr IWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateStack(const StackQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateStack(const StackQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateStridedSlice(const StridedSliceQueueDescriptor& descriptor, - const WorkloadInfo& Info) const +std::unique_ptr IWorkloadFactory::CreateStridedSlice(const StridedSliceQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateSubtraction(const SubtractionQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateSubtraction(const SubtractionQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } -std::unique_ptr IWorkloadFactory::CreateSwitch(const SwitchQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr IWorkloadFactory::CreateSwitch(const SwitchQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } std::unique_ptr IWorkloadFactory::CreateTransposeConvolution2d( - const TransposeConvolution2dQueueDescriptor& descriptor, - const WorkloadInfo& info) const + const TransposeConvolution2dQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return std::unique_ptr(); } diff --git a/src/backends/backendsCommon/WorkloadFactoryBase.hpp b/src/backends/backendsCommon/WorkloadFactoryBase.hpp index 6f7437413a..1947c6935b 100644 --- a/src/backends/backendsCommon/WorkloadFactoryBase.hpp +++ b/src/backends/backendsCommon/WorkloadFactoryBase.hpp @@ -16,240 +16,242 @@ public: bool SupportsSubTensors() const override { return false; }; - std::unique_ptr CreateSubTensorHandle(ITensorHandle& parent, TensorShape const& subTensorShape, - unsigned int const *subTensorOrigin) const override + std::unique_ptr CreateSubTensorHandle(ITensorHandle& /*parent*/, + TensorShape const& /*subTensorShape*/, + unsigned int const */*subTensorOrigin*/) const override { return nullptr; }; - std::unique_ptr CreateInput(const InputQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateInput(const InputQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateTensorHandle(const TensorInfo& tensorInfo, - const bool IsMemoryManaged = true) const override + std::unique_ptr CreateTensorHandle(const TensorInfo& /*tensorInfo*/, + const bool /*IsMemoryManaged*/) const override { return nullptr; } - std::unique_ptr CreateTensorHandle(const TensorInfo& tensorInfo, DataLayout dataLayout, - const bool IsMemoryManaged = true) const override + std::unique_ptr CreateTensorHandle(const TensorInfo& /*tensorInfo*/, + DataLayout /*dataLayout*/, + const bool /*IsMemoryManaged*/) const override { return nullptr; } - std::unique_ptr CreateAbs(const AbsQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateAbs(const AbsQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateActivation(const ActivationQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateActivation(const ActivationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateAddition(const AdditionQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateAddition(const AdditionQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateArgMinMax(const ArgMinMaxQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateArgMinMax(const ArgMinMaxQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateBatchNormalization(const BatchNormalizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateBatchNormalization(const BatchNormalizationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& descriptor, - const WorkloadInfo& Info) const override + std::unique_ptr CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateComparison(const ComparisonQueueDescriptor& descriptor, - const WorkloadInfo& Info) const override + std::unique_ptr CreateComparison(const ComparisonQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateConcat(const ConcatQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateConcat(const ConcatQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateConstant(const ConstantQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateConstant(const ConstantQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateConvertFp16ToFp32(const ConvertFp16ToFp32QueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateConvertFp16ToFp32(const ConvertFp16ToFp32QueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateConvertFp32ToFp16(const ConvertFp32ToFp16QueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateConvertFp32ToFp16(const ConvertFp32ToFp16QueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateConvolution2d(const Convolution2dQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateDebug(const DebugQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateDebug(const DebugQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateDepthToSpace(const DepthToSpaceQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateDepthToSpace(const DepthToSpaceQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateDepthwiseConvolution2d(const DepthwiseConvolution2dQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateDepthwiseConvolution2d(const DepthwiseConvolution2dQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateDequantize(const DequantizeQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateDequantize(const DequantizeQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateDetectionPostProcess(const DetectionPostProcessQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateDetectionPostProcess(const DetectionPostProcessQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateDivision(const DivisionQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateDivision(const DivisionQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateFakeQuantization(const FakeQuantizationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateFloor(const FloorQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateFloor(const FloorQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateFullyConnected(const FullyConnectedQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateFullyConnected(const FullyConnectedQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateGather(const GatherQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateGather(const GatherQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateInstanceNormalization(const InstanceNormalizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateInstanceNormalization(const InstanceNormalizationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateL2Normalization(const L2NormalizationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateLogSoftmax(const LogSoftmaxQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateLogSoftmax(const LogSoftmaxQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateLstm(const LstmQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateLstm(const LstmQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateMaximum(const MaximumQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateMaximum(const MaximumQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateMean(const MeanQueueDescriptor& descriptor, - const WorkloadInfo& Info) const override + std::unique_ptr CreateMean(const MeanQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateMemCopy(const MemCopyQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateMemCopy(const MemCopyQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateMemImport(const MemImportQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateMemImport(const MemImportQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateMerge(const MergeQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateMerge(const MergeQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateMinimum(const MinimumQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateMinimum(const MinimumQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateMultiplication(const MultiplicationQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateMultiplication(const MultiplicationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateNormalization(const NormalizationQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateNormalization(const NormalizationQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateOutput(const OutputQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateOutput(const OutputQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreatePad(const PadQueueDescriptor& descriptor, - const WorkloadInfo& Info) const override + std::unique_ptr CreatePad(const PadQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreatePermute(const PermuteQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreatePermute(const PermuteQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreatePooling2d(const Pooling2dQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreatePooling2d(const Pooling2dQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreatePreCompiled(const PreCompiledQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreatePreCompiled(const PreCompiledQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreatePrelu(const PreluQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreatePrelu(const PreluQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateQuantize(const QuantizeQueueDescriptor& descriptor, - const WorkloadInfo& Info) const override + std::unique_ptr CreateQuantize(const QuantizeQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateQuantizedLstm(const QuantizedLstmQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateQuantizedLstm(const QuantizedLstmQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateReshape(const ReshapeQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateReshape(const ReshapeQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateResize(const ResizeQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateResize(const ResizeQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateRsqrt(const RsqrtQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateRsqrt(const RsqrtQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateSlice(const SliceQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateSlice(const SliceQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateSoftmax(const SoftmaxQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateSoftmax(const SoftmaxQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateSubtraction(const SubtractionQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateSubtraction(const SubtractionQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateSplitter(const SplitterQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateSplitter(const SplitterQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateStack(const StackQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateStack(const StackQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateStridedSlice(const StridedSliceQueueDescriptor& descriptor, - const WorkloadInfo& Info) const override + std::unique_ptr CreateStridedSlice(const StridedSliceQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateSwitch(const SwitchQueueDescriptor& descriptor, - const WorkloadInfo& Info) const override + std::unique_ptr CreateSwitch(const SwitchQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr CreateTransposeConvolution2d(const TransposeConvolution2dQueueDescriptor& descriptor, - const WorkloadInfo& info) const override + std::unique_ptr CreateTransposeConvolution2d(const TransposeConvolution2dQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const override { return nullptr; } }; diff --git a/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp b/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp index 7ab5ee4ec4..6924beb820 100644 --- a/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp +++ b/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp @@ -414,6 +414,7 @@ struct LayerTypePolicy \ static std::unique_ptr MakeDummyWorkload(armnn::IWorkloadFactory *factory, \ unsigned int nIn, unsigned int nOut) \ { \ + boost::ignore_unused(factory, nIn, nOut); \ return std::unique_ptr(); \ } \ }; diff --git a/src/backends/backendsCommon/test/MockBackend.cpp b/src/backends/backendsCommon/test/MockBackend.cpp index c5a4ed9e52..367d9cb8a5 100644 --- a/src/backends/backendsCommon/test/MockBackend.cpp +++ b/src/backends/backendsCommon/test/MockBackend.cpp @@ -88,7 +88,7 @@ const BackendId& MockBackend::GetIdStatic() } IBackendInternal::IWorkloadFactoryPtr MockBackend::CreateWorkloadFactory( - const IBackendInternal::IMemoryManagerSharedPtr& memoryManager) const + const IBackendInternal::IMemoryManagerSharedPtr& /*memoryManager*/) const { return IWorkloadFactoryPtr{}; } diff --git a/src/backends/backendsCommon/test/MockBackend.hpp b/src/backends/backendsCommon/test/MockBackend.hpp index 437f23d2b3..771e499280 100644 --- a/src/backends/backendsCommon/test/MockBackend.hpp +++ b/src/backends/backendsCommon/test/MockBackend.hpp @@ -37,32 +37,32 @@ public: class MockLayerSupport : public LayerSupportBase { public: - bool IsInputSupported(const TensorInfo& input, - Optional reasonIfUnsupported = EmptyOptional()) const override + bool IsInputSupported(const TensorInfo& /*input*/, + Optional /*reasonIfUnsupported = EmptyOptional()*/) const override { return true; } - bool IsOutputSupported(const TensorInfo& input, - Optional reasonIfUnsupported = EmptyOptional()) const override + bool IsOutputSupported(const TensorInfo& /*input*/, + Optional /*reasonIfUnsupported = EmptyOptional()*/) const override { return true; } - bool IsAdditionSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, - Optional reasonIfUnsupported = EmptyOptional()) const override + bool IsAdditionSupported(const TensorInfo& /*input0*/, + const TensorInfo& /*input1*/, + const TensorInfo& /*output*/, + Optional /*reasonIfUnsupported = EmptyOptional()*/) const override { return true; } - bool IsConvolution2dSupported(const TensorInfo& input, - const TensorInfo& output, - const Convolution2dDescriptor& descriptor, - const TensorInfo& weights, - const Optional& biases, - Optional reasonIfUnsupported = EmptyOptional()) const override + bool IsConvolution2dSupported(const TensorInfo& /*input*/, + const TensorInfo& /*output*/, + const Convolution2dDescriptor& /*descriptor*/, + const TensorInfo& /*weights*/, + const Optional& /*biases*/, + Optional /*reasonIfUnsupported = EmptyOptional()*/) const override { return true; } diff --git a/src/backends/backendsCommon/test/TestDynamicBackend.cpp b/src/backends/backendsCommon/test/TestDynamicBackend.cpp index a53f1698c2..cbfe09377e 100644 --- a/src/backends/backendsCommon/test/TestDynamicBackend.cpp +++ b/src/backends/backendsCommon/test/TestDynamicBackend.cpp @@ -7,6 +7,8 @@ #include +#include + constexpr const char* TestDynamicBackendId() { #if defined(VALID_TEST_DYNAMIC_BACKEND_1) @@ -63,6 +65,7 @@ public: } IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager) const override { + boost::ignore_unused(memoryManager); return IWorkloadFactoryPtr{}; } ILayerSupportSharedPtr GetLayerSupport() const override diff --git a/src/backends/cl/ClLayerSupport.cpp b/src/backends/cl/ClLayerSupport.cpp index 6e1e9d98c6..bce91ab462 100644 --- a/src/backends/cl/ClLayerSupport.cpp +++ b/src/backends/cl/ClLayerSupport.cpp @@ -88,8 +88,10 @@ bool IsMatchingStride(uint32_t actualStride) return IsMatchingStride(actualStride) || IsMatchingStride(actualStride); } -bool IsClBackendSupported(Optional reasonIfUnsupported) +template +bool IsClBackendSupported(Optional reasonIfUnsupported, Args... args) { + boost::ignore_unused(reasonIfUnsupported, (args)...); #if defined(ARMCOMPUTECL_ENABLED) return true; #else @@ -124,7 +126,7 @@ inline bool IsWorkloadSupported(FuncType&& func, Optional reasonIf return IsWorkloadSupported(func, reasonIfUnsupported, __VA_ARGS__); #else #define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \ - return IsClBackendSupported(reasonIfUnsupported); + return IsClBackendSupported(reasonIfUnsupported, __VA_ARGS__); #endif template @@ -461,7 +463,7 @@ bool ClLayerSupport::IsGreaterSupported(const TensorInfo& input0, bool ClLayerSupport::IsInputSupported(const TensorInfo& input, Optional reasonIfUnsupported) const { - return IsClBackendSupported(reasonIfUnsupported); + return IsClBackendSupported(reasonIfUnsupported, input); } bool ClLayerSupport::IsInstanceNormalizationSupported(const TensorInfo& input, @@ -579,7 +581,7 @@ bool ClLayerSupport::IsNormalizationSupported(const TensorInfo& input, bool ClLayerSupport::IsOutputSupported(const TensorInfo& output, Optional reasonIfUnsupported) const { - return IsClBackendSupported(reasonIfUnsupported); + return IsClBackendSupported(reasonIfUnsupported, output); } bool ClLayerSupport::IsPadSupported(const TensorInfo& input, @@ -758,6 +760,7 @@ bool ClLayerSupport::IsSplitterSupported(const TensorInfo& input, *splitAxis.begin()); } #endif + boost::ignore_unused(descriptor); for (auto output : outputs) { if (!input.IsTypeSpaceMatch(output)) // Cannot use sub-tensors if the types are not same space diff --git a/src/backends/cl/ClWorkloadFactory.cpp b/src/backends/cl/ClWorkloadFactory.cpp index 531f3710ea..4746167795 100644 --- a/src/backends/cl/ClWorkloadFactory.cpp +++ b/src/backends/cl/ClWorkloadFactory.cpp @@ -23,6 +23,7 @@ #include #include +#include #include #include @@ -84,6 +85,7 @@ ClWorkloadFactory::ClWorkloadFactory(const std::shared_ptr& mem std::unique_ptr ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo, const bool IsMemoryManaged) const { + boost::ignore_unused(IsMemoryManaged); std::unique_ptr tensorHandle = std::make_unique(tensorInfo); tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup()); @@ -94,6 +96,7 @@ std::unique_ptr ClWorkloadFactory::CreateTensorHandle(const Tenso DataLayout dataLayout, const bool IsMemoryManaged) const { + boost::ignore_unused(IsMemoryManaged); std::unique_ptr tensorHandle = std::make_unique(tensorInfo, dataLayout); tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup()); diff --git a/src/backends/cl/OpenClTimer.cpp b/src/backends/cl/OpenClTimer.cpp index 57552d7bd9..ee3c114ba0 100644 --- a/src/backends/cl/OpenClTimer.cpp +++ b/src/backends/cl/OpenClTimer.cpp @@ -8,6 +8,8 @@ #include #include +#include + namespace armnn { @@ -29,6 +31,7 @@ void OpenClTimer::Start() const cl_event * event_wait_list, cl_event * event) { + boost::ignore_unused(event); cl_int retVal = 0; // Get the name of the kernel diff --git a/src/backends/neon/NeonLayerSupport.cpp b/src/backends/neon/NeonLayerSupport.cpp index 3fc323393e..c79aa78d1b 100644 --- a/src/backends/neon/NeonLayerSupport.cpp +++ b/src/backends/neon/NeonLayerSupport.cpp @@ -66,8 +66,10 @@ namespace armnn namespace { -bool IsNeonBackendSupported(Optional reasonIfUnsupported) +template< typename ... Args> +bool IsNeonBackendSupported(Optional reasonIfUnsupported, Args... args) { + boost::ignore_unused((args)...); #if defined(ARMCOMPUTENEON_ENABLED) return true; #else @@ -111,7 +113,7 @@ inline bool IsWorkloadSupported(FuncType& func, Optional reasonIfU return IsWorkloadSupported(func, reasonIfUnsupported, __VA_ARGS__); #else #define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \ - return IsNeonBackendSupported(reasonIfUnsupported); + return IsNeonBackendSupported(reasonIfUnsupported, __VA_ARGS__); #endif #if defined(ARMCOMPUTENEON_ENABLED) @@ -427,7 +429,7 @@ bool NeonLayerSupport::IsGreaterSupported(const armnn::TensorInfo& input0, bool NeonLayerSupport::IsInputSupported(const TensorInfo& input, Optional reasonIfUnsupported) const { - return IsNeonBackendSupported(reasonIfUnsupported); + return IsNeonBackendSupported(reasonIfUnsupported, input); } bool NeonLayerSupport::IsInstanceNormalizationSupported(const TensorInfo& input, @@ -545,7 +547,7 @@ bool NeonLayerSupport::IsNormalizationSupported(const TensorInfo& input, bool NeonLayerSupport::IsOutputSupported(const TensorInfo& output, Optional reasonIfUnsupported) const { - return IsNeonBackendSupported(reasonIfUnsupported); + return IsNeonBackendSupported(reasonIfUnsupported, output); } bool NeonLayerSupport::IsPadSupported(const TensorInfo& input, @@ -721,6 +723,7 @@ bool NeonLayerSupport::IsSplitterSupported(const TensorInfo& input, *splitAxis.begin()); } #endif + boost::ignore_unused(descriptor); for (auto output : outputs) { if (!input.IsTypeSpaceMatch(output)) // Cannot use sub-tensors if the types are not same space diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp index 19b76152f3..ebcd1f633e 100644 --- a/src/backends/reference/RefLayerSupport.cpp +++ b/src/backends/reference/RefLayerSupport.cpp @@ -648,6 +648,8 @@ bool RefLayerSupport::IsDetectionPostProcessSupported(const TensorInfo& boxEncod const DetectionPostProcessDescriptor& descriptor, Optional reasonIfUnsupported) const { + boost::ignore_unused(anchors, detectionBoxes, detectionClasses, detectionScores, numDetections, descriptor); + bool supported = true; std::array supportedInputTypes = @@ -863,8 +865,8 @@ bool RefLayerSupport::IsGreaterSupported(const TensorInfo& input0, reasonIfUnsupported); } -bool RefLayerSupport::IsInputSupported(const TensorInfo& input, - Optional reasonIfUnsupported) const +bool RefLayerSupport::IsInputSupported(const TensorInfo& /*input*/, + Optional /*reasonIfUnsupported*/) const { return true; } @@ -1301,8 +1303,8 @@ bool RefLayerSupport::IsNormalizationSupported(const TensorInfo& input, return supported; } -bool RefLayerSupport::IsOutputSupported(const TensorInfo& output, - Optional reasonIfUnsupported) const +bool RefLayerSupport::IsOutputSupported(const TensorInfo& /*output*/, + Optional /*reasonIfUnsupported*/) const { return true; } @@ -1470,6 +1472,7 @@ bool RefLayerSupport::IsResizeSupported(const TensorInfo& input, const ResizeDescriptor& descriptor, Optional reasonIfUnsupported) const { + boost::ignore_unused(descriptor); bool supported = true; std::array supportedTypes = { @@ -1524,7 +1527,7 @@ bool RefLayerSupport::IsSliceSupported(const TensorInfo& input, const SliceDescriptor& descriptor, Optional reasonIfUnsupported) const { - ignore_unused(descriptor); + boost::ignore_unused(descriptor); bool supported = true; std::array supportedTypes = @@ -1551,7 +1554,7 @@ bool RefLayerSupport::IsSoftmaxSupported(const TensorInfo& input, const SoftmaxDescriptor& descriptor, Optional reasonIfUnsupported) const { - ignore_unused(output); + boost::ignore_unused(descriptor); bool supported = true; std::array supportedTypes = { @@ -1578,7 +1581,7 @@ bool RefLayerSupport::IsSpaceToBatchNdSupported(const TensorInfo& input, const SpaceToBatchNdDescriptor& descriptor, Optional reasonIfUnsupported) const { - ignore_unused(output); + boost::ignore_unused(descriptor); bool supported = true; std::array supportedTypes = { @@ -1811,6 +1814,7 @@ bool RefLayerSupport::IsTransposeConvolution2dSupported(const TensorInfo& input, const Optional& biases, Optional reasonIfUnsupported) const { + boost::ignore_unused(descriptor); bool supported = true; std::array supportedTypes = diff --git a/src/backends/reference/RefWorkloadFactory.cpp b/src/backends/reference/RefWorkloadFactory.cpp index 8d044eecb7..dffb13db2d 100644 --- a/src/backends/reference/RefWorkloadFactory.cpp +++ b/src/backends/reference/RefWorkloadFactory.cpp @@ -77,19 +77,21 @@ bool RefWorkloadFactory::IsLayerSupported(const Layer& layer, } std::unique_ptr RefWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo, - const bool IsMemoryManaged) const + const bool isMemoryManaged) const { // For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer // to unmanaged memory. This also ensures memory alignment. + boost::ignore_unused(isMemoryManaged); return std::make_unique(tensorInfo, m_MemoryManager); } std::unique_ptr RefWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo, DataLayout dataLayout, - const bool IsMemoryManaged) const + const bool isMemoryManaged) const { // For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer // to unmanaged memory. This also ensures memory alignment. + boost::ignore_unused(isMemoryManaged, dataLayout); return std::make_unique(tensorInfo, m_MemoryManager); } @@ -218,6 +220,7 @@ std::unique_ptr RefWorkloadFactory::CreateDivision(const DivisionQueu std::unique_ptr RefWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor, const WorkloadInfo& info) const { + boost::ignore_unused(descriptor); ComparisonQueueDescriptor comparisonDescriptor; comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Equal; @@ -253,6 +256,7 @@ std::unique_ptr RefWorkloadFactory::CreateGather(const GatherQueueDes std::unique_ptr RefWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor, const WorkloadInfo& info) const { + boost::ignore_unused(descriptor); ComparisonQueueDescriptor comparisonDescriptor; comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Greater; @@ -410,8 +414,8 @@ std::unique_ptr RefWorkloadFactory::CreatePooling2d(const Pooling2dQu return std::make_unique(descriptor, info); } -std::unique_ptr RefWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& descriptor, - const WorkloadInfo& info) const +std::unique_ptr RefWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& /*descriptor*/, + const WorkloadInfo& /*info*/) const { return nullptr; } diff --git a/src/backends/reference/test/RefWorkloadFactoryHelper.hpp b/src/backends/reference/test/RefWorkloadFactoryHelper.hpp index b49a6dd3d5..10e5b9fa28 100644 --- a/src/backends/reference/test/RefWorkloadFactoryHelper.hpp +++ b/src/backends/reference/test/RefWorkloadFactoryHelper.hpp @@ -25,6 +25,7 @@ struct WorkloadFactoryHelper static armnn::RefWorkloadFactory GetFactory( const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager = nullptr) { + boost::ignore_unused(memoryManager); return armnn::RefWorkloadFactory(); } }; diff --git a/src/backends/reference/workloads/ArgMinMax.cpp b/src/backends/reference/workloads/ArgMinMax.cpp index 76616f1465..db85b958e9 100644 --- a/src/backends/reference/workloads/ArgMinMax.cpp +++ b/src/backends/reference/workloads/ArgMinMax.cpp @@ -15,6 +15,8 @@ namespace armnn void ArgMinMax(Decoder& in, int32_t* out, const TensorInfo& inputTensorInfo, const TensorInfo& outputTensorInfo, ArgMinMaxFunction function, int axis) { + boost::ignore_unused(outputTensorInfo); + unsigned int uAxis = armnnUtils::GetUnsignedAxis(inputTensorInfo.GetNumDimensions(), axis); const unsigned int outerElements = armnnUtils::GetNumElementsBetween(inputTensorInfo.GetShape(), 0, uAxis); diff --git a/src/backends/reference/workloads/Dequantize.cpp b/src/backends/reference/workloads/Dequantize.cpp index fafc03e69b..4025e8d7fa 100644 --- a/src/backends/reference/workloads/Dequantize.cpp +++ b/src/backends/reference/workloads/Dequantize.cpp @@ -5,6 +5,7 @@ #include "Dequantize.hpp" +#include namespace armnn { @@ -13,6 +14,7 @@ void Dequantize(Decoder& inputDecoder, const TensorInfo& inputInfo, const TensorInfo& outputInfo) { + boost::ignore_unused(outputInfo); BOOST_ASSERT(inputInfo.GetNumElements() == outputInfo.GetNumElements()); for (unsigned int i = 0; i < inputInfo.GetNumElements(); i++) { diff --git a/src/backends/reference/workloads/DetectionPostProcess.cpp b/src/backends/reference/workloads/DetectionPostProcess.cpp index d475dd8ac0..3b384f1480 100644 --- a/src/backends/reference/workloads/DetectionPostProcess.cpp +++ b/src/backends/reference/workloads/DetectionPostProcess.cpp @@ -155,6 +155,8 @@ void DetectionPostProcess(const TensorInfo& boxEncodingsInfo, float* detectionScores, float* numDetections) { + boost::ignore_unused(anchorsInfo, detectionClassesInfo, detectionScoresInfo, numDetectionsInfo); + // Transform center-size format which is (ycenter, xcenter, height, width) to box-corner format, // which represents the lower left corner and the upper right corner (ymin, xmin, ymax, xmax) std::vector boxCorners(boxEncodingsInfo.GetNumElements()); diff --git a/src/backends/reference/workloads/Gather.cpp b/src/backends/reference/workloads/Gather.cpp index c848a7c138..5416855f48 100644 --- a/src/backends/reference/workloads/Gather.cpp +++ b/src/backends/reference/workloads/Gather.cpp @@ -9,6 +9,7 @@ #include +#include #include namespace armnn @@ -21,6 +22,7 @@ void Gather(const TensorInfo& paramsInfo, const int32_t* indices, Encoder& output) { + boost::ignore_unused(outputInfo); const TensorShape& paramsShape = paramsInfo.GetShape(); unsigned int paramsProduct = 1; diff --git a/src/backends/reference/workloads/Pooling2d.cpp b/src/backends/reference/workloads/Pooling2d.cpp index ea8f4ee7e7..8ff2eb457a 100644 --- a/src/backends/reference/workloads/Pooling2d.cpp +++ b/src/backends/reference/workloads/Pooling2d.cpp @@ -84,7 +84,7 @@ namespace { case PoolingAlgorithm::Max: { - return [](float & accumulated, float kernelSize) {}; + return [](float & /*accumulated*/, float /*kernelSize*/) {}; } case PoolingAlgorithm::Average: -- cgit v1.2.1