diff options
Diffstat (limited to 'src/backends/backendsCommon')
-rw-r--r-- | src/backends/backendsCommon/CMakeLists.txt | 1 | ||||
-rw-r--r-- | src/backends/backendsCommon/IBackendInternal.cpp | 41 | ||||
-rw-r--r-- | src/backends/backendsCommon/IBackendInternal.hpp | 9 | ||||
-rw-r--r-- | src/backends/backendsCommon/LayerSupportBase.cpp | 78 | ||||
-rw-r--r-- | src/backends/backendsCommon/LayerSupportBase.hpp | 44 | ||||
-rw-r--r-- | src/backends/backendsCommon/TensorHandle.hpp | 21 | ||||
-rw-r--r-- | src/backends/backendsCommon/WorkloadData.cpp | 10 | ||||
-rw-r--r-- | src/backends/backendsCommon/WorkloadData.hpp | 7 | ||||
-rw-r--r-- | src/backends/backendsCommon/WorkloadFactory.cpp | 37 | ||||
-rw-r--r-- | src/backends/backendsCommon/WorkloadFactory.hpp | 24 | ||||
-rw-r--r-- | src/backends/backendsCommon/WorkloadFactoryBase.hpp | 16 | ||||
-rw-r--r-- | src/backends/backendsCommon/WorkloadUtils.cpp | 4 | ||||
-rw-r--r-- | src/backends/backendsCommon/test/MockBackend.cpp | 5 | ||||
-rw-r--r-- | src/backends/backendsCommon/test/MockBackend.hpp | 1 | ||||
-rw-r--r-- | src/backends/backendsCommon/test/mockBackend/MockImportBackend.cpp | 5 | ||||
-rw-r--r-- | src/backends/backendsCommon/test/mockBackend/MockImportBackend.hpp | 1 |
16 files changed, 27 insertions, 277 deletions
diff --git a/src/backends/backendsCommon/CMakeLists.txt b/src/backends/backendsCommon/CMakeLists.txt index 129cdbe9f1..c894f986c9 100644 --- a/src/backends/backendsCommon/CMakeLists.txt +++ b/src/backends/backendsCommon/CMakeLists.txt @@ -12,7 +12,6 @@ list(APPEND armnnBackendsCommon_sources DynamicBackendUtils.hpp IBackendContext.hpp IBackendInternal.cpp - IBackendInternal.hpp IMemoryManager.hpp ITensorHandle.hpp ITensorHandleFactory.cpp diff --git a/src/backends/backendsCommon/IBackendInternal.cpp b/src/backends/backendsCommon/IBackendInternal.cpp index 31706eb1e7..ec1313df0c 100644 --- a/src/backends/backendsCommon/IBackendInternal.cpp +++ b/src/backends/backendsCommon/IBackendInternal.cpp @@ -9,26 +9,6 @@ namespace armnn { -ARMNN_NO_DEPRECATE_WARN_BEGIN -IBackendInternal::ISubGraphConverterPtr IBackendInternal::CreateSubGraphConverter( - const std::shared_ptr<SubGraph>& /*subGrapg*/) const -{ - return ISubGraphConverterPtr{}; -} - -IBackendInternal::Optimizations IBackendInternal::GetOptimizations() const -{ - return Optimizations{}; -} - -IBackendInternal::SubGraphUniquePtr IBackendInternal::OptimizeSubGraph(const SubGraph& /*subGraph*/, - bool& optimizationAttempted) const -{ - optimizationAttempted = false; - return nullptr; -} -ARMNN_NO_DEPRECATE_WARN_END - IMemoryManagerUniquePtr IBackendInternal::CreateMemoryManager() const { return IMemoryManagerUniquePtr(); @@ -120,29 +100,12 @@ IBackendInternal::ILayerSupportSharedPtr IBackendInternal::GetLayerSupport(const return GetLayerSupport(); } -// Default implementation of OptimizeSubgraphView for backward compatibility with the old API. +// Default implementation of OptimizeSubgraphView. Returns an untouched subgraph. // Override this method with a custom optimization implementation. OptimizationViews IBackendInternal::OptimizeSubgraphView(const SubgraphView& subgraph) const { - bool optimizationAttempted = false; - - ARMNN_NO_DEPRECATE_WARN_BEGIN - SubGraphUniquePtr optSubgraph = OptimizeSubGraph(subgraph, optimizationAttempted); - ARMNN_NO_DEPRECATE_WARN_END - OptimizationViews result; - if (!optimizationAttempted) - { - result.AddUntouchedSubgraph(SubgraphView(subgraph)); - } - else if (optSubgraph) - { - result.AddSubstitution({subgraph, SubgraphView(*optSubgraph.get())}); - } - else - { - result.AddFailedSubgraph(SubgraphView(subgraph)); - } + result.AddUntouchedSubgraph(SubgraphView(subgraph)); return result; } diff --git a/src/backends/backendsCommon/IBackendInternal.hpp b/src/backends/backendsCommon/IBackendInternal.hpp deleted file mode 100644 index 61ccc4f057..0000000000 --- a/src/backends/backendsCommon/IBackendInternal.hpp +++ /dev/null @@ -1,9 +0,0 @@ -// -// Copyright © 2019 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -// This file is depricated and will be removed soon. -// Please use the new header in armnn/backends instead. -// This will use the new armnn/backends header. -#include <armnn/backends/IBackendInternal.hpp> diff --git a/src/backends/backendsCommon/LayerSupportBase.cpp b/src/backends/backendsCommon/LayerSupportBase.cpp index 2c3f827622..ca1acc376b 100644 --- a/src/backends/backendsCommon/LayerSupportBase.cpp +++ b/src/backends/backendsCommon/LayerSupportBase.cpp @@ -37,13 +37,6 @@ bool DefaultLayerSupport(const char* func, namespace armnn { -bool LayerSupportBase::IsAbsSupported(const TensorInfo&, // input - const TensorInfo&, // output - Optional<std::string &> reasonIfUnsupported) const -{ - return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); -} - bool LayerSupportBase::IsActivationSupported(const TensorInfo&, // input const TensorInfo&, //output const ActivationDescriptor&, // descriptor @@ -238,31 +231,11 @@ bool LayerSupportBase::IsDivisionSupported(const TensorInfo&, // input0 return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsElementwiseUnarySupported(const TensorInfo& input, - const TensorInfo& output, - const ElementwiseUnaryDescriptor& descriptor, +bool LayerSupportBase::IsElementwiseUnarySupported(const TensorInfo&, // input + const TensorInfo&, // output + const ElementwiseUnaryDescriptor&, // descriptor Optional<std::string&> reasonIfUnsupported) const { - if (descriptor.m_Operation == UnaryOperation::Abs) - { - ARMNN_NO_DEPRECATE_WARN_BEGIN - return IsAbsSupported(input, output, reasonIfUnsupported); - ARMNN_NO_DEPRECATE_WARN_END - } - else if (descriptor.m_Operation == UnaryOperation::Rsqrt) - { - ARMNN_NO_DEPRECATE_WARN_BEGIN - return IsRsqrtSupported(input, output, reasonIfUnsupported); - ARMNN_NO_DEPRECATE_WARN_END - } - return false; -} - -bool LayerSupportBase::IsEqualSupported(const armnn::TensorInfo&, // input0 - const armnn::TensorInfo&, // input1 - const armnn::TensorInfo&, // output - armnn::Optional<std::string &> reasonIfUnsupported) const -{ return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } @@ -301,28 +274,12 @@ bool LayerSupportBase::IsFullyConnectedSupported(const TensorInfo&, // input bool LayerSupportBase::IsGatherSupported(const armnn::TensorInfo&, // input0 const armnn::TensorInfo&, // input1 const armnn::TensorInfo&, // output - armnn::Optional<std::string&> reasonIfUnsupported) const -{ - return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); -} - -bool LayerSupportBase::IsGatherSupported(const armnn::TensorInfo&, // input0 - const armnn::TensorInfo&, // input1 - const armnn::TensorInfo&, // output const GatherDescriptor&, // descriptor armnn::Optional<std::string&> reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsGreaterSupported(const TensorInfo&, // input0 - const TensorInfo&, // input1 - const TensorInfo&, // output - Optional<std::string&> reasonIfUnsupported) const -{ - return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); -} - bool LayerSupportBase::IsInputSupported(const TensorInfo&, // input Optional<std::string&> reasonIfUnsupported) const { @@ -422,14 +379,6 @@ bool LayerSupportBase::IsMergeSupported(const TensorInfo&, // input0 return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsMergerSupported(const std::vector<const TensorInfo*> inputs, - const TensorInfo& output, - const OriginsDescriptor& descriptor, - Optional<std::string&> reasonIfUnsupported) const -{ - return IsConcatSupported(inputs, output, descriptor, reasonIfUnsupported); -} - bool LayerSupportBase::IsMinimumSupported(const TensorInfo&, // input0 const TensorInfo&, // input1 const TensorInfo&, // output @@ -553,13 +502,6 @@ bool LayerSupportBase::IsReshapeSupported(const TensorInfo&, // input return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsResizeBilinearSupported(const TensorInfo&, // input - const TensorInfo&, // output - Optional<std::string&> reasonIfUnsupported) const -{ - return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); -} - bool LayerSupportBase::IsResizeSupported(const TensorInfo&, // input const TensorInfo&, // output const ResizeDescriptor&, // descriptor @@ -568,13 +510,6 @@ bool LayerSupportBase::IsResizeSupported(const TensorInfo&, // input return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); } -bool LayerSupportBase::IsRsqrtSupported(const TensorInfo&, // input - const TensorInfo&, // output - Optional<std::string &> reasonIfUnsupported) const -{ - return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); -} - bool LayerSupportBase::IsShapeSupported(const TensorInfo&, // input const TensorInfo&, // output Optional<std::string&> reasonIfUnsupported) const @@ -615,13 +550,6 @@ bool LayerSupportBase::IsSpaceToDepthSupported(const TensorInfo&, // input } bool LayerSupportBase::IsSplitterSupported(const TensorInfo&, // input - const ViewsDescriptor&, // descriptor - Optional<std::string&> reasonIfUnsupported) const -{ - return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); -} - -bool LayerSupportBase::IsSplitterSupported(const TensorInfo&, // input const std::vector<std::reference_wrapper<TensorInfo>>&, // outputs const ViewsDescriptor&, // descriptor Optional<std::string&> reasonIfUnsupported) const diff --git a/src/backends/backendsCommon/LayerSupportBase.hpp b/src/backends/backendsCommon/LayerSupportBase.hpp index 240b1dab73..fc2906f497 100644 --- a/src/backends/backendsCommon/LayerSupportBase.hpp +++ b/src/backends/backendsCommon/LayerSupportBase.hpp @@ -13,11 +13,6 @@ namespace armnn class LayerSupportBase : public ILayerSupport { public: - ARMNN_DEPRECATED_MSG("Use IsElementwiseUnarySupported instead") - bool IsAbsSupported(const TensorInfo& input, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsActivationSupported(const TensorInfo& input, const TensorInfo& output, const ActivationDescriptor& descriptor, @@ -149,12 +144,6 @@ public: const ElementwiseUnaryDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsComparisonSupported instead") - bool IsEqualSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsFakeQuantizationSupported(const TensorInfo& input, const FakeQuantizationDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; @@ -175,24 +164,12 @@ public: const FullyConnectedDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsGatherSupported with descriptor instead") - bool IsGatherSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsGatherSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, const GatherDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsComparisonSupported instead") - bool IsGreaterSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsInputSupported(const TensorInfo& input, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; @@ -257,12 +234,6 @@ public: const TensorInfo& output, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsConcatSupported instead") - bool IsMergerSupported(const std::vector<const TensorInfo*> inputs, - const TensorInfo& output, - const OriginsDescriptor& descriptor, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsMinimumSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, @@ -346,16 +317,6 @@ public: const ResizeDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsResizeSupported instead") - bool IsResizeBilinearSupported(const TensorInfo& input, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - - ARMNN_DEPRECATED_MSG("Use IsElementwiseUnarySupported instead") - bool IsRsqrtSupported(const TensorInfo& input, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsShapeSupported(const TensorInfo& input, const TensorInfo& output, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; @@ -380,11 +341,6 @@ public: const SpaceToDepthDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsSplitterSupported with outputs instead") - bool IsSplitterSupported(const TensorInfo& input, - const ViewsDescriptor& descriptor, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsSplitterSupported(const TensorInfo& input, const std::vector<std::reference_wrapper<TensorInfo>>& outputs, const ViewsDescriptor& descriptor, diff --git a/src/backends/backendsCommon/TensorHandle.hpp b/src/backends/backendsCommon/TensorHandle.hpp index 4e9d87d6eb..b898bd11a5 100644 --- a/src/backends/backendsCommon/TensorHandle.hpp +++ b/src/backends/backendsCommon/TensorHandle.hpp @@ -242,16 +242,17 @@ private: std::shared_ptr<ConstTensorHandle> m_TensorHandle; }; -using ConstCpuTensorHandle ARMNN_DEPRECATED_MSG("ConstCpuTensorHandle is deprecated, " - "use ConstTensorHandle instead") = ConstTensorHandle; -using CpuTensorHandle ARMNN_DEPRECATED_MSG("CpuTensorHandle is deprecated, " - "use TensorHandle instead") = TensorHandle; -using ScopedCpuTensorHandle ARMNN_DEPRECATED_MSG("ScopedCpuTensorHandle is deprecated, " - "use ScopedTensorHandle instead") = ScopedTensorHandle; -using PassthroughCpuTensorHandle ARMNN_DEPRECATED_MSG("PassthroughCpuTensorHandle is deprecated, use " - "PassthroughTensorHandle instead") = PassthroughTensorHandle; -using ConstPassthroughCpuTensorHandle ARMNN_DEPRECATED_MSG("ConstPassthroughCpuTensorHandle is " +using ConstCpuTensorHandle ARMNN_DEPRECATED_MSG_REMOVAL_DATE("ConstCpuTensorHandle is deprecated, " + "use ConstTensorHandle instead", "22.05") = ConstTensorHandle; +using CpuTensorHandle ARMNN_DEPRECATED_MSG_REMOVAL_DATE("CpuTensorHandle is deprecated, " + "use TensorHandle instead", "22.05") = TensorHandle; +using ScopedCpuTensorHandle ARMNN_DEPRECATED_MSG_REMOVAL_DATE("ScopedCpuTensorHandle is deprecated, " + "use ScopedTensorHandle instead", "22.05") = ScopedTensorHandle; +using PassthroughCpuTensorHandle ARMNN_DEPRECATED_MSG_REMOVAL_DATE("PassthroughCpuTensorHandle is deprecated, use " + "PassthroughTensorHandle instead", + "22.05") = PassthroughTensorHandle; +using ConstPassthroughCpuTensorHandle ARMNN_DEPRECATED_MSG_REMOVAL_DATE("ConstPassthroughCpuTensorHandle is " "deprecated, use ConstPassthroughTensorHandle " - "instead") = ConstPassthroughTensorHandle; + "instead", "22.05") = ConstPassthroughTensorHandle; } // namespace armnn diff --git a/src/backends/backendsCommon/WorkloadData.cpp b/src/backends/backendsCommon/WorkloadData.cpp index fe22133104..27b59ea3a6 100644 --- a/src/backends/backendsCommon/WorkloadData.cpp +++ b/src/backends/backendsCommon/WorkloadData.cpp @@ -157,15 +157,12 @@ void ValidateTensorDataType(const TensorInfo& tensor, DataType dataType, void ValidPerAxisQuantizedDataType(const TensorInfo& tensor, const std::string& descName, const std::string& tensorName) { - ARMNN_NO_DEPRECATE_WARN_BEGIN - if (tensor.GetDataType() != DataType::QSymmS8 && - tensor.GetDataType() != DataType::QuantizedSymm8PerAxis) + if (tensor.GetDataType() != DataType::QSymmS8) { throw InvalidArgumentException(descName + ": Expected data type which supports per-axis quantization scheme but got " + GetDataTypeName(tensor.GetDataType()) + " for " + tensorName + " tensor."); } - ARMNN_NO_DEPRECATE_WARN_END } //--------------------------------------------------------------- @@ -362,15 +359,12 @@ void ValidateWeightDataType(const TensorInfo& inputInfo, const DataType inputType = inputInfo.GetDataType(); if (IsQuantized8BitType(inputType)) { - ARMNN_NO_DEPRECATE_WARN_BEGIN const std::vector<DataType> validTypes = { DataType::QAsymmS8, DataType::QAsymmU8, - DataType::QSymmS8, - DataType::QuantizedSymm8PerAxis // deprecated + DataType::QSymmS8 }; - ARMNN_NO_DEPRECATE_WARN_END ValidateDataTypes(weightInfo, validTypes, descName); } diff --git a/src/backends/backendsCommon/WorkloadData.hpp b/src/backends/backendsCommon/WorkloadData.hpp index 896081ecfd..29d39d14a9 100644 --- a/src/backends/backendsCommon/WorkloadData.hpp +++ b/src/backends/backendsCommon/WorkloadData.hpp @@ -345,10 +345,15 @@ struct RankQueueDescriptor : QueueDescriptor void Validate(const WorkloadInfo& workloadInfo) const; }; -struct ResizeBilinearQueueDescriptor : QueueDescriptorWithParameters<ResizeBilinearDescriptor> +ARMNN_NO_DEPRECATE_WARN_BEGIN +struct +ARMNN_DEPRECATED_MSG_REMOVAL_DATE("ResizeBilinearQueueDescriptor is deprecated use ResizeQueueDescriptor instead", + "22.08") +ResizeBilinearQueueDescriptor : QueueDescriptorWithParameters<ResizeBilinearDescriptor> { void Validate(const WorkloadInfo& workloadInfo) const; }; +ARMNN_NO_DEPRECATE_WARN_END struct ResizeQueueDescriptor : QueueDescriptorWithParameters<ResizeDescriptor> { diff --git a/src/backends/backendsCommon/WorkloadFactory.cpp b/src/backends/backendsCommon/WorkloadFactory.cpp index 666f83de71..3b7f3a0f1f 100644 --- a/src/backends/backendsCommon/WorkloadFactory.cpp +++ b/src/backends/backendsCommon/WorkloadFactory.cpp @@ -1499,13 +1499,6 @@ bool IWorkloadFactory::IsLayerSupported(const BackendId& backendId, modelOptions); } -// Default Implementations -std::unique_ptr<IWorkload> IWorkloadFactory::CreateAbs(const AbsQueueDescriptor& /*descriptor*/, - const WorkloadInfo& /*info*/) const -{ - return std::unique_ptr<IWorkload>(); -} - std::unique_ptr<IWorkload> IWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const { @@ -1644,12 +1637,6 @@ std::unique_ptr<IWorkload> IWorkloadFactory::CreateElementwiseUnary(const Elemen return std::unique_ptr<IWorkload>(); } -std::unique_ptr<IWorkload> IWorkloadFactory::CreateEqual(const EqualQueueDescriptor& /*descriptor*/, - const WorkloadInfo& /*Info*/) const -{ - return std::unique_ptr<IWorkload>(); -} - std::unique_ptr<IWorkload> IWorkloadFactory::CreateFakeQuantization(const FakeQuantizationQueueDescriptor& /*desc*/, const WorkloadInfo& /*info*/) const { @@ -1680,12 +1667,6 @@ std::unique_ptr<IWorkload> IWorkloadFactory::CreateGather(const GatherQueueDescr return std::unique_ptr<IWorkload>(); } -std::unique_ptr<IWorkload> IWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& /*descriptor*/, - const WorkloadInfo& /*info*/) const -{ - return std::unique_ptr<IWorkload>(); -} - std::unique_ptr<IWorkload> IWorkloadFactory::CreateInstanceNormalization( const InstanceNormalizationQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const @@ -1753,12 +1734,6 @@ std::unique_ptr<IWorkload> IWorkloadFactory::CreateMerge(const MergeQueueDescrip return std::unique_ptr<IWorkload>(); } -std::unique_ptr<IWorkload> IWorkloadFactory::CreateMerger(const MergerQueueDescriptor& /*descriptor*/, - const WorkloadInfo& /*info*/) const -{ - return std::unique_ptr<IWorkload>(); -} - std::unique_ptr<IWorkload> IWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const { @@ -1848,24 +1823,12 @@ std::unique_ptr<IWorkload> IWorkloadFactory::CreateReshape(const ReshapeQueueDes return std::unique_ptr<IWorkload>(); } -std::unique_ptr<IWorkload> IWorkloadFactory::CreateResizeBilinear(const ResizeBilinearQueueDescriptor& /*descriptor*/, - const WorkloadInfo& /*info*/) const -{ - return std::unique_ptr<IWorkload>(); -} - std::unique_ptr<IWorkload> IWorkloadFactory::CreateResize(const ResizeQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const { return std::unique_ptr<IWorkload>(); } -std::unique_ptr<IWorkload> IWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& /*descriptor*/, - const WorkloadInfo& /*info*/) const -{ - return std::unique_ptr<IWorkload>(); -} - std::unique_ptr<IWorkload> IWorkloadFactory::CreateShape(const ShapeQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const { diff --git a/src/backends/backendsCommon/WorkloadFactory.hpp b/src/backends/backendsCommon/WorkloadFactory.hpp index c16fcb882b..df4bcd6144 100644 --- a/src/backends/backendsCommon/WorkloadFactory.hpp +++ b/src/backends/backendsCommon/WorkloadFactory.hpp @@ -68,10 +68,6 @@ public: DataLayout dataLayout, const bool IsMemoryManaged = true) const = 0; - ARMNN_DEPRECATED_MSG("Use CreateElementwiseUnary instead") - virtual std::unique_ptr<IWorkload> CreateAbs(const AbsQueueDescriptor& descriptor, - const WorkloadInfo& info) const; - virtual std::unique_ptr<IWorkload> CreateActivation(const ActivationQueueDescriptor& descriptor, const WorkloadInfo& info) const; @@ -141,10 +137,6 @@ public: virtual std::unique_ptr<IWorkload> CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& descriptor, const WorkloadInfo& Info) const; - ARMNN_DEPRECATED_MSG("Use CreateComparison instead") - virtual std::unique_ptr<IWorkload> CreateEqual(const EqualQueueDescriptor& descriptor, - const WorkloadInfo& Info) const; - virtual std::unique_ptr<IWorkload> CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor, const WorkloadInfo& info) const; @@ -160,10 +152,6 @@ public: virtual std::unique_ptr<IWorkload> CreateGather(const GatherQueueDescriptor& descriptor, const WorkloadInfo& info) const; - ARMNN_DEPRECATED_MSG("Use CreateComparison instead") - virtual std::unique_ptr<IWorkload> CreateGreater(const GreaterQueueDescriptor& descriptor, - const WorkloadInfo& info) const; - virtual std::unique_ptr<IWorkload> CreateInstanceNormalization( const InstanceNormalizationQueueDescriptor& descriptor, const WorkloadInfo& info) const; @@ -198,10 +186,6 @@ public: virtual std::unique_ptr<IWorkload> CreateMerge(const MergeQueueDescriptor& descriptor, const WorkloadInfo& info) const; - ARMNN_DEPRECATED_MSG("Use CreateConcat instead") - virtual std::unique_ptr<IWorkload> CreateMerger(const MergerQueueDescriptor& descriptor, - const WorkloadInfo& info) const; - virtual std::unique_ptr<IWorkload> CreateMinimum(const MinimumQueueDescriptor& descriptor, const WorkloadInfo& info) const; @@ -250,14 +234,6 @@ public: virtual std::unique_ptr<IWorkload> CreateResize(const ResizeQueueDescriptor& descriptor, const WorkloadInfo& info) const; - ARMNN_DEPRECATED_MSG("Use CreateResize instead") - virtual std::unique_ptr<IWorkload> CreateResizeBilinear(const ResizeBilinearQueueDescriptor& descriptor, - const WorkloadInfo& info) const; - - ARMNN_DEPRECATED_MSG("Use CreateElementwiseUnary instead") - virtual std::unique_ptr<IWorkload> CreateRsqrt(const RsqrtQueueDescriptor& descriptor, - const WorkloadInfo& info) const; - virtual std::unique_ptr<IWorkload> CreateShape(const ShapeQueueDescriptor& descriptor, const WorkloadInfo& info) const; diff --git a/src/backends/backendsCommon/WorkloadFactoryBase.hpp b/src/backends/backendsCommon/WorkloadFactoryBase.hpp index 295202324e..ef507a64f8 100644 --- a/src/backends/backendsCommon/WorkloadFactoryBase.hpp +++ b/src/backends/backendsCommon/WorkloadFactoryBase.hpp @@ -34,10 +34,6 @@ public: const bool /*IsMemoryManaged*/) const override { return nullptr; } - std::unique_ptr<IWorkload> CreateAbs(const AbsQueueDescriptor& /*descriptor*/, - const WorkloadInfo& /*info*/) const override - { return nullptr; } - std::unique_ptr<IWorkload> CreateActivation(const ActivationQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const override { return nullptr; } @@ -111,19 +107,17 @@ public: { if (descriptor.m_Parameters.m_Operation == UnaryOperation::Abs) { - AbsQueueDescriptor absDescriptor; - return CreateAbs(absDescriptor, info); + { return nullptr; } } else if (descriptor.m_Parameters.m_Operation == UnaryOperation::Rsqrt) { - RsqrtQueueDescriptor rsqrtDescriptor; - return CreateRsqrt(rsqrtDescriptor, info); + { return nullptr; } } else if (descriptor.m_Parameters.m_Operation == UnaryOperation::LogicalNot) { return CreateLogicalUnary(descriptor, info); } - return nullptr; + { return nullptr; } } std::unique_ptr<IWorkload> CreateFakeQuantization(const FakeQuantizationQueueDescriptor& /*descriptor*/, @@ -234,10 +228,6 @@ public: const WorkloadInfo& /*info*/) const override { return nullptr; } - std::unique_ptr<IWorkload> CreateRsqrt(const RsqrtQueueDescriptor& /*descriptor*/, - const WorkloadInfo& /*info*/) const override - { return nullptr; } - std::unique_ptr<IWorkload> CreateSlice(const SliceQueueDescriptor& /*descriptor*/, const WorkloadInfo& /*info*/) const override { return nullptr; } diff --git a/src/backends/backendsCommon/WorkloadUtils.cpp b/src/backends/backendsCommon/WorkloadUtils.cpp index bd7f09b28a..fe681936f1 100644 --- a/src/backends/backendsCommon/WorkloadUtils.cpp +++ b/src/backends/backendsCommon/WorkloadUtils.cpp @@ -265,13 +265,9 @@ armnn::ConstTensor ConvertWeightTensorFromArmnnToAcl(const ConstTensorHandle* we case DataType::QAsymmU8: weightPermuted = ReorderWeightChannelsForAcl<uint8_t>(weightPermuted, dataLayout, permuteBuffer); break; - ARMNN_NO_DEPRECATE_WARN_BEGIN - case DataType::QuantizedSymm8PerAxis: - ARMNN_FALLTHROUGH; case DataType::QSymmS8: weightPermuted = ReorderWeightChannelsForAcl<int8_t>(weightPermuted, dataLayout, permuteBuffer); break; - ARMNN_NO_DEPRECATE_WARN_END default: break; } diff --git a/src/backends/backendsCommon/test/MockBackend.cpp b/src/backends/backendsCommon/test/MockBackend.cpp index e706fc8157..df1a5c19aa 100644 --- a/src/backends/backendsCommon/test/MockBackend.cpp +++ b/src/backends/backendsCommon/test/MockBackend.cpp @@ -117,11 +117,6 @@ IBackendInternal::IMemoryManagerUniquePtr MockBackend::CreateMemoryManager() con return IMemoryManagerUniquePtr{}; } -IBackendInternal::Optimizations MockBackend::GetOptimizations() const -{ - return Optimizations{}; -} - IBackendInternal::ILayerSupportSharedPtr MockBackend::GetLayerSupport() const { static ILayerSupportSharedPtr layerSupport{new MockLayerSupport}; diff --git a/src/backends/backendsCommon/test/MockBackend.hpp b/src/backends/backendsCommon/test/MockBackend.hpp index d90ad798da..c0624525dc 100644 --- a/src/backends/backendsCommon/test/MockBackend.hpp +++ b/src/backends/backendsCommon/test/MockBackend.hpp @@ -162,7 +162,6 @@ public: CreateBackendProfilingContext(const IRuntime::CreationOptions& creationOptions, IBackendProfilingPtr& backendProfiling) override; - IBackendInternal::Optimizations GetOptimizations() const override; IBackendInternal::ILayerSupportSharedPtr GetLayerSupport() const override; OptimizationViews OptimizeSubgraphView(const SubgraphView& subgraph) const override; diff --git a/src/backends/backendsCommon/test/mockBackend/MockImportBackend.cpp b/src/backends/backendsCommon/test/mockBackend/MockImportBackend.cpp index ebe94348fc..ea6ece7b32 100644 --- a/src/backends/backendsCommon/test/mockBackend/MockImportBackend.cpp +++ b/src/backends/backendsCommon/test/mockBackend/MockImportBackend.cpp @@ -79,11 +79,6 @@ IBackendInternal::IMemoryManagerUniquePtr MockImportBackend::CreateMemoryManager return std::make_unique<RefMemoryManager>(); } -IBackendInternal::Optimizations MockImportBackend::GetOptimizations() const -{ - return Optimizations{}; -} - IBackendInternal::ILayerSupportSharedPtr MockImportBackend::GetLayerSupport() const { static ILayerSupportSharedPtr layerSupport{new MockImportLayerSupport}; diff --git a/src/backends/backendsCommon/test/mockBackend/MockImportBackend.hpp b/src/backends/backendsCommon/test/mockBackend/MockImportBackend.hpp index ecc661f43b..c07a97c29e 100644 --- a/src/backends/backendsCommon/test/mockBackend/MockImportBackend.hpp +++ b/src/backends/backendsCommon/test/mockBackend/MockImportBackend.hpp @@ -40,7 +40,6 @@ public: IBackendInternal::IBackendProfilingContextPtr CreateBackendProfilingContext( const IRuntime::CreationOptions& creationOptions, IBackendProfilingPtr& backendProfiling) override; - IBackendInternal::Optimizations GetOptimizations() const override; IBackendInternal::ILayerSupportSharedPtr GetLayerSupport() const override; OptimizationViews OptimizeSubgraphView(const SubgraphView& subgraph) const override; |