diff options
Diffstat (limited to 'src/backends/reference')
-rw-r--r-- | src/backends/reference/RefBackend.cpp | 15 | ||||
-rw-r--r-- | src/backends/reference/RefBackend.hpp | 3 | ||||
-rw-r--r-- | src/backends/reference/RefLayerSupport.cpp | 119 | ||||
-rw-r--r-- | src/backends/reference/RefLayerSupport.hpp | 37 | ||||
-rw-r--r-- | src/backends/reference/RefWorkloadFactory.cpp | 58 | ||||
-rw-r--r-- | src/backends/reference/RefWorkloadFactory.hpp | 24 | ||||
-rw-r--r-- | src/backends/reference/backend.mk | 1 | ||||
-rw-r--r-- | src/backends/reference/workloads/CMakeLists.txt | 2 | ||||
-rw-r--r-- | src/backends/reference/workloads/Decoders.hpp | 7 | ||||
-rw-r--r-- | src/backends/reference/workloads/Encoders.hpp | 10 | ||||
-rw-r--r-- | src/backends/reference/workloads/RefResizeBilinearWorkload.cpp | 45 | ||||
-rw-r--r-- | src/backends/reference/workloads/RefResizeBilinearWorkload.hpp | 24 | ||||
-rw-r--r-- | src/backends/reference/workloads/RefWorkloads.hpp | 1 |
13 files changed, 5 insertions, 341 deletions
diff --git a/src/backends/reference/RefBackend.cpp b/src/backends/reference/RefBackend.cpp index c9f164e0c9..a3060f0798 100644 --- a/src/backends/reference/RefBackend.cpp +++ b/src/backends/reference/RefBackend.cpp @@ -58,27 +58,12 @@ IBackendInternal::IMemoryManagerUniquePtr RefBackend::CreateMemoryManager() cons return std::make_unique<RefMemoryManager>(); } -IBackendInternal::Optimizations RefBackend::GetOptimizations() const -{ - return Optimizations{}; -} - IBackendInternal::ILayerSupportSharedPtr RefBackend::GetLayerSupport() const { static ILayerSupportSharedPtr layerSupport{new RefLayerSupport}; return layerSupport; } -bool RefBackend::HasCapability(BackendCapability capabilityClass) const -{ - auto search = oldCpuRefCapabilities.find(capabilityClass); - if (search != oldCpuRefCapabilities.end()) - { - return true; - } - return false; -} - OptimizationViews RefBackend::OptimizeSubgraphView(const SubgraphView& subgraph) const { OptimizationViews optimizationViews; diff --git a/src/backends/reference/RefBackend.hpp b/src/backends/reference/RefBackend.hpp index 28c1591179..c04bf43db3 100644 --- a/src/backends/reference/RefBackend.hpp +++ b/src/backends/reference/RefBackend.hpp @@ -47,7 +47,6 @@ public: IBackendInternal::IBackendProfilingContextPtr CreateBackendProfilingContext( const IRuntime::CreationOptions& creationOptions, IBackendProfilingPtr& backendProfiling) override; - IBackendInternal::Optimizations GetOptimizations() const override; IBackendInternal::ILayerSupportSharedPtr GetLayerSupport() const override; OptimizationViews OptimizeSubgraphView(const SubgraphView& subgraph) const override; @@ -60,8 +59,6 @@ public: { return cpuRefCapabilities; }; - - bool HasCapability(BackendCapability capabilityClass) const override; }; } // namespace armnn diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp index c0ede678bf..b80aa9992f 100644 --- a/src/backends/reference/RefLayerSupport.cpp +++ b/src/backends/reference/RefLayerSupport.cpp @@ -58,15 +58,6 @@ std::string CreateIncorrectDimensionsErrorMsg(unsigned int expected, } // anonymous namespace -bool RefLayerSupport::IsAbsSupported(const TensorInfo& input, const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported) const -{ - return IsElementwiseUnarySupported(input, - output, - ElementwiseUnaryDescriptor(UnaryOperation::Abs), - reasonIfUnsupported); -} - bool RefLayerSupport::IsActivationSupported(const TensorInfo& input, const TensorInfo& output, const ActivationDescriptor& descriptor, @@ -565,15 +556,12 @@ bool RefLayerSupport::IsConvolution2dSupported(const TensorInfo& input, const DataType inputType = input.GetDataType(); if (IsQuantized8BitType(inputType)) { - ARMNN_NO_DEPRECATE_WARN_BEGIN - std::array<DataType, 4> supportedWeightTypes = + std::array<DataType, 3> supportedWeightTypes = { DataType::QAsymmS8, DataType::QAsymmU8, - DataType::QSymmS8, - DataType::QuantizedSymm8PerAxis // deprecated + DataType::QSymmS8 }; - ARMNN_NO_DEPRECATE_WARN_END supported &= CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported, "Reference Convolution2d: weights type not supported for quantized input."); @@ -769,15 +757,12 @@ bool RefLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input, const DataType inputType = input.GetDataType(); if (IsQuantized8BitType(inputType)) { - ARMNN_NO_DEPRECATE_WARN_BEGIN - std::array<DataType, 4> supportedWeightTypes = + std::array<DataType, 3> supportedWeightTypes = { DataType::QAsymmS8, DataType::QAsymmU8, DataType::QSymmS8, - DataType::QuantizedSymm8PerAxis // deprecated }; - ARMNN_NO_DEPRECATE_WARN_END supported &= CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported, "Reference DepthwiseConvolution2d: weights type not supported for " @@ -977,18 +962,6 @@ bool RefLayerSupport::IsElementwiseUnarySupported(const TensorInfo& input, return supported; } -bool RefLayerSupport::IsEqualSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported) const -{ - return IsComparisonSupported(input0, - input1, - output, - ComparisonDescriptor(ComparisonOperation::Equal), - reasonIfUnsupported); -} - bool RefLayerSupport::IsFakeQuantizationSupported(const TensorInfo& input, const FakeQuantizationDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const @@ -1173,18 +1146,6 @@ bool RefLayerSupport::IsGatherSupported(const armnn::TensorInfo& input0, return supported; } -bool RefLayerSupport::IsGreaterSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported) const -{ - return IsComparisonSupported(input0, - input1, - output, - ComparisonDescriptor(ComparisonOperation::Greater), - reasonIfUnsupported); -} - bool RefLayerSupport::IsInputSupported(const TensorInfo& /*input*/, Optional<std::string&> /*reasonIfUnsupported*/) const { @@ -1523,14 +1484,6 @@ bool RefLayerSupport::IsMeanSupported(const TensorInfo& input, return supported; } -bool RefLayerSupport::IsMergerSupported(const std::vector<const TensorInfo*> inputs, - const TensorInfo& output, - const MergerDescriptor& descriptor, - Optional<std::string&> reasonIfUnsupported) const -{ - return IsConcatSupported(inputs, output, descriptor, reasonIfUnsupported); -} - bool RefLayerSupport::IsMemCopySupported(const TensorInfo &input, const TensorInfo &output, Optional<std::string &> reasonIfUnsupported) const @@ -1897,33 +1850,6 @@ bool RefLayerSupport::IsReshapeSupported(const TensorInfo& input, "Reference reshape: input type not supported."); } -bool RefLayerSupport::IsResizeBilinearSupported(const TensorInfo& input, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported) const -{ - bool supported = true; - std::array<DataType,6> supportedTypes = - { - DataType::BFloat16, - DataType::Float32, - DataType::Float16, - DataType::QAsymmS8, - DataType::QAsymmU8, - DataType::QSymmS16 - }; - - supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, - "Reference ResizeBilinear: input type not supported"); - - supported &= CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported, - "Reference ResizeBilinear: output type not supported"); - - supported &= CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported, - "Reference ResizeBilinear: input and output types not matching"); - - return supported; -} - bool RefLayerSupport::IsResizeSupported(const TensorInfo& input, const TensorInfo& output, const ResizeDescriptor& descriptor, @@ -1953,16 +1879,6 @@ bool RefLayerSupport::IsResizeSupported(const TensorInfo& input, return supported; } -bool RefLayerSupport::IsRsqrtSupported(const TensorInfo& input, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported) const -{ - return IsElementwiseUnarySupported(input, - output, - ElementwiseUnaryDescriptor(UnaryOperation::Rsqrt), - reasonIfUnsupported); -} - bool RefLayerSupport::IsShapeSupported(const TensorInfo& input, const TensorInfo& output, Optional<std::string&> reasonIfUnsupported) const @@ -2101,28 +2017,6 @@ bool RefLayerSupport::IsSpaceToDepthSupported(const TensorInfo& input, } bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input, - const ViewsDescriptor& descriptor, - Optional<std::string&> reasonIfUnsupported) const -{ - IgnoreUnused(descriptor); - bool supported = true; - std::array<DataType,6> supportedTypes = - { - DataType::BFloat16, - DataType::Float32, - DataType::Float16, - DataType::QAsymmS8, - DataType::QAsymmU8, - DataType::QSymmS16 - }; - - supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, - "Reference splitter: input type not supported"); - - return supported; -} - -bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input, const std::vector<std::reference_wrapper<TensorInfo>>& outputs, const ViewsDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const @@ -2322,15 +2216,12 @@ bool RefLayerSupport::IsTransposeConvolution2dSupported(const TensorInfo& input, const DataType inputType = input.GetDataType(); if (IsQuantized8BitType(inputType)) { - ARMNN_NO_DEPRECATE_WARN_BEGIN - std::array<DataType, 4> supportedWeightTypes = + std::array<DataType, 3> supportedWeightTypes = { DataType::QAsymmS8, DataType::QAsymmU8, - DataType::QSymmS8, - DataType::QuantizedSymm8PerAxis //Deprecated + DataType::QSymmS8 }; - ARMNN_NO_DEPRECATE_WARN_END supported &= CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported, "Reference TransposeConvolution2d: weights type not supported for " diff --git a/src/backends/reference/RefLayerSupport.hpp b/src/backends/reference/RefLayerSupport.hpp index 627418e3e1..53d7907204 100644 --- a/src/backends/reference/RefLayerSupport.hpp +++ b/src/backends/reference/RefLayerSupport.hpp @@ -12,11 +12,6 @@ namespace armnn class RefLayerSupport : public LayerSupportBase { public: - ARMNN_DEPRECATED_MSG("Use IsElementwiseUnarySupported instead") - bool IsAbsSupported(const TensorInfo& input, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsActivationSupported(const TensorInfo& input, const TensorInfo& output, const ActivationDescriptor& descriptor, @@ -147,12 +142,6 @@ public: const ElementwiseUnaryDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsComparisonSupported instead") - bool IsEqualSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsFakeQuantizationSupported(const TensorInfo& input, const FakeQuantizationDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; @@ -179,12 +168,6 @@ public: const GatherDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsComparisonSupported instead") - bool IsGreaterSupported(const TensorInfo& input0, - const TensorInfo& input1, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsInputSupported(const TensorInfo& input, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; @@ -230,12 +213,6 @@ public: const MeanDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsConcatSupported instead") - bool IsMergerSupported(const std::vector<const TensorInfo*> inputs, - const TensorInfo& output, - const MergerDescriptor& descriptor, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsMemCopySupported(const TensorInfo& input, const TensorInfo& output, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; @@ -301,20 +278,11 @@ public: const ReshapeDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsResizeBilinearSupported(const TensorInfo& input, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsResizeSupported(const TensorInfo& input, const TensorInfo& output, const ResizeDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsElementwiseUnarySupported instead") - bool IsRsqrtSupported(const TensorInfo& input, - const TensorInfo& output, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsShapeSupported(const TensorInfo& input, const TensorInfo& output, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; @@ -340,11 +308,6 @@ public: Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - ARMNN_DEPRECATED_MSG("Use IsSplitterSupported with outputs instead") - bool IsSplitterSupported(const TensorInfo& input, - const ViewsDescriptor& descriptor, - Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; - bool IsSplitterSupported(const TensorInfo& input, const std::vector<std::reference_wrapper<TensorInfo>>& outputs, const ViewsDescriptor& descriptor, diff --git a/src/backends/reference/RefWorkloadFactory.cpp b/src/backends/reference/RefWorkloadFactory.cpp index 18a5af277f..75008bc866 100644 --- a/src/backends/reference/RefWorkloadFactory.cpp +++ b/src/backends/reference/RefWorkloadFactory.cpp @@ -129,16 +129,6 @@ std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const Tens return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager); } -std::unique_ptr<IWorkload> RefWorkloadFactory::CreateAbs(const AbsQueueDescriptor& descriptor, - const WorkloadInfo& info) const -{ - IgnoreUnused(descriptor); - ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor; - elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Abs; - - return CreateElementwiseUnary(elementwiseUnaryDescriptor, info); -} - std::unique_ptr<IWorkload> RefWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor, const WorkloadInfo& info) const { @@ -331,16 +321,6 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateElementwiseUnary(const Elem return std::make_unique<RefElementwiseUnaryWorkload>(descriptor, info); } -std::unique_ptr<IWorkload> RefWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor, - const WorkloadInfo& info) const -{ - IgnoreUnused(descriptor); - ComparisonQueueDescriptor comparisonDescriptor; - comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Equal; - - return CreateComparison(comparisonDescriptor, info); -} - std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor, const WorkloadInfo& info) const { @@ -379,16 +359,6 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGather(const GatherQueueDes return std::make_unique<RefGatherWorkload>(descriptor, info); } -std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor, - const WorkloadInfo& info) const -{ - IgnoreUnused(descriptor); - ComparisonQueueDescriptor comparisonDescriptor; - comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Greater; - - return CreateComparison(comparisonDescriptor, info); -} - std::unique_ptr<IWorkload> RefWorkloadFactory::CreateInput(const InputQueueDescriptor& descriptor, const WorkloadInfo& info) const { @@ -479,12 +449,6 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMemImport(const MemImportQu return std::make_unique<ImportMemGenericWorkload>(descriptor, info); } -std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMerger(const MergerQueueDescriptor& descriptor, - const WorkloadInfo& info) const -{ - return CreateConcat(descriptor, info); -} - std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& descriptor, const WorkloadInfo& info) const { @@ -615,28 +579,6 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResize(const ResizeQueueDes return std::make_unique<RefResizeWorkload>(descriptor, info); } -std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResizeBilinear(const ResizeBilinearQueueDescriptor& descriptor, - const WorkloadInfo& info) const -{ - ResizeQueueDescriptor resizeDescriptor; - resizeDescriptor.m_Parameters.m_Method = ResizeMethod::Bilinear; - resizeDescriptor.m_Parameters.m_DataLayout = descriptor.m_Parameters.m_DataLayout; - resizeDescriptor.m_Parameters.m_TargetWidth = descriptor.m_Parameters.m_TargetWidth; - resizeDescriptor.m_Parameters.m_TargetHeight = descriptor.m_Parameters.m_TargetHeight; - - return CreateResize(resizeDescriptor, info); -} - -std::unique_ptr<IWorkload> RefWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& descriptor, - const WorkloadInfo& info) const -{ - IgnoreUnused(descriptor); - ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor; - elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Rsqrt; - - return CreateElementwiseUnary(elementwiseUnaryDescriptor, info); -} - std::unique_ptr<IWorkload> RefWorkloadFactory::CreateShape(const ShapeQueueDescriptor& descriptor, const WorkloadInfo& info) const { diff --git a/src/backends/reference/RefWorkloadFactory.hpp b/src/backends/reference/RefWorkloadFactory.hpp index d00d3ca822..a85e8dda3e 100644 --- a/src/backends/reference/RefWorkloadFactory.hpp +++ b/src/backends/reference/RefWorkloadFactory.hpp @@ -66,10 +66,6 @@ public: DataLayout dataLayout, const bool IsMemoryManaged = true) const override; - ARMNN_DEPRECATED_MSG("Use CreateElementwiseUnary instead") - std::unique_ptr<IWorkload> CreateAbs(const AbsQueueDescriptor& descriptor, - const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateActivation(const ActivationQueueDescriptor& descriptor, const WorkloadInfo& info) const override; @@ -139,10 +135,6 @@ public: std::unique_ptr<IWorkload> CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& descriptor, const WorkloadInfo& info) const override; - ARMNN_DEPRECATED_MSG("Use CreateComparison instead") - std::unique_ptr<IWorkload> CreateEqual(const EqualQueueDescriptor& descriptor, - const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor, const WorkloadInfo& info) const override; @@ -158,10 +150,6 @@ public: std::unique_ptr<IWorkload> CreateGather(const GatherQueueDescriptor& descriptor, const WorkloadInfo& info) const override; - ARMNN_DEPRECATED_MSG("Use CreateComparison instead") - std::unique_ptr<IWorkload> CreateGreater(const GreaterQueueDescriptor& descriptor, - const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateInput(const InputQueueDescriptor& descriptor, const WorkloadInfo& info) const override; @@ -192,10 +180,6 @@ public: std::unique_ptr<IWorkload> CreateMemImport(const MemImportQueueDescriptor& descriptor, const WorkloadInfo& info) const override; - ARMNN_DEPRECATED_MSG("Use CreateConcat instead") - std::unique_ptr<IWorkload> CreateMerger(const MergerQueueDescriptor& descriptor, - const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateMinimum(const MinimumQueueDescriptor& descriptor, const WorkloadInfo& info) const override; @@ -241,14 +225,6 @@ public: std::unique_ptr<IWorkload> CreateResize(const ResizeQueueDescriptor& descriptor, const WorkloadInfo& info) const override; - ARMNN_DEPRECATED_MSG("Use CreateResize instead") - std::unique_ptr<IWorkload> CreateResizeBilinear(const ResizeBilinearQueueDescriptor& descriptor, - const WorkloadInfo& info) const override; - - ARMNN_DEPRECATED_MSG("Use CreateElementwiseUnary instead") - std::unique_ptr<IWorkload> CreateRsqrt(const RsqrtQueueDescriptor& descriptor, - const WorkloadInfo& info) const override; - std::unique_ptr<IWorkload> CreateShape(const ShapeQueueDescriptor& descriptor, const WorkloadInfo& info) const override; diff --git a/src/backends/reference/backend.mk b/src/backends/reference/backend.mk index 7d6c59a273..f8169a6c0c 100644 --- a/src/backends/reference/backend.mk +++ b/src/backends/reference/backend.mk @@ -88,7 +88,6 @@ BACKEND_SOURCES := \ workloads/RefQuantizeWorkload.cpp \ workloads/RefReduceWorkload.cpp \ workloads/RefReshapeWorkload.cpp \ - workloads/RefResizeBilinearWorkload.cpp \ workloads/RefResizeWorkload.cpp \ workloads/RefSliceWorkload.cpp \ workloads/RefSoftmaxWorkload.cpp \ diff --git a/src/backends/reference/workloads/CMakeLists.txt b/src/backends/reference/workloads/CMakeLists.txt index e169c03ad8..5727291be3 100644 --- a/src/backends/reference/workloads/CMakeLists.txt +++ b/src/backends/reference/workloads/CMakeLists.txt @@ -148,8 +148,6 @@ list(APPEND armnnRefBackendWorkloads_sources RefReduceWorkload.hpp RefReshapeWorkload.cpp RefReshapeWorkload.hpp - RefResizeBilinearWorkload.cpp - RefResizeBilinearWorkload.hpp RefResizeWorkload.cpp RefResizeWorkload.hpp RefShapeWorkload.hpp diff --git a/src/backends/reference/workloads/Decoders.hpp b/src/backends/reference/workloads/Decoders.hpp index cd0dc5d40f..c2a456bfce 100644 --- a/src/backends/reference/workloads/Decoders.hpp +++ b/src/backends/reference/workloads/Decoders.hpp @@ -67,13 +67,6 @@ inline std::unique_ptr<Decoder<float>> MakeDecoder(const TensorInfo& info, const { switch(info.GetDataType()) { - ARMNN_NO_DEPRECATE_WARN_BEGIN - case armnn::DataType::QuantizedSymm8PerAxis: - { - std::pair<unsigned int, std::vector<float>> params = armnnUtils::GetPerAxisParams(info); - return std::make_unique<QSymm8PerAxisDecoder>(static_cast<const int8_t*>(data), info); - } - ARMNN_NO_DEPRECATE_WARN_END case DataType::QAsymmS8: { return std::make_unique<QASymmS8Decoder>( diff --git a/src/backends/reference/workloads/Encoders.hpp b/src/backends/reference/workloads/Encoders.hpp index a2d565ec4a..a7be9e172b 100644 --- a/src/backends/reference/workloads/Encoders.hpp +++ b/src/backends/reference/workloads/Encoders.hpp @@ -22,16 +22,6 @@ inline std::unique_ptr<Encoder<float>> MakeEncoder(const TensorInfo& info, void* { switch(info.GetDataType()) { - ARMNN_NO_DEPRECATE_WARN_BEGIN - case armnn::DataType::QuantizedSymm8PerAxis: - { - std::pair<unsigned int, std::vector<float>> params = armnnUtils::GetPerAxisParams(info); - return std::make_unique<QSymm8PerAxisEncoder>( - static_cast<int8_t*>(data), - params.second, - params.first); - } - ARMNN_NO_DEPRECATE_WARN_END case armnn::DataType::QAsymmS8: { return std::make_unique<QASymmS8Encoder>( diff --git a/src/backends/reference/workloads/RefResizeBilinearWorkload.cpp b/src/backends/reference/workloads/RefResizeBilinearWorkload.cpp deleted file mode 100644 index 2cf5888f33..0000000000 --- a/src/backends/reference/workloads/RefResizeBilinearWorkload.cpp +++ /dev/null @@ -1,45 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#include "RefResizeBilinearWorkload.hpp" - -#include "RefWorkloadUtils.hpp" -#include "Resize.hpp" -#include "BaseIterator.hpp" -#include "Profiling.hpp" - -#include "BaseIterator.hpp" -#include "Decoders.hpp" -#include "Encoders.hpp" - -namespace armnn -{ - -void RefResizeBilinearWorkload::Execute() const -{ - Execute(m_Data.m_Inputs, m_Data.m_Outputs); -} - -void RefResizeBilinearWorkload::ExecuteAsync(WorkingMemDescriptor &workingMemDescriptor) -{ - Execute(workingMemDescriptor.m_Inputs, workingMemDescriptor.m_Outputs); -} - -void RefResizeBilinearWorkload::Execute(std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs) const -{ - ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefResizeBilinearWorkload_Execute"); - - const TensorInfo& inputInfo = GetTensorInfo(inputs[0]); - const TensorInfo& outputInfo = GetTensorInfo(outputs[0]); - - std::unique_ptr<Decoder<float>> decoderPtr = MakeDecoder<float>(inputInfo, inputs[0]->Map()); - Decoder<float> &decoder = *decoderPtr; - std::unique_ptr<Encoder<float>> encoderPtr = MakeEncoder<float>(outputInfo, outputs[0]->Map()); - Encoder<float> &encoder = *encoderPtr; - - Resize(decoder, inputInfo, encoder, outputInfo, m_Data.m_Parameters.m_DataLayout, armnn::ResizeMethod::Bilinear); -} - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefResizeBilinearWorkload.hpp b/src/backends/reference/workloads/RefResizeBilinearWorkload.hpp deleted file mode 100644 index 5ada3d1ff8..0000000000 --- a/src/backends/reference/workloads/RefResizeBilinearWorkload.hpp +++ /dev/null @@ -1,24 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#pragma once - -#include <backendsCommon/Workload.hpp> -#include <backendsCommon/WorkloadData.hpp> - -namespace armnn -{ - -class RefResizeBilinearWorkload : public BaseWorkload<ResizeBilinearQueueDescriptor> -{ -public: - using BaseWorkload<ResizeBilinearQueueDescriptor>::BaseWorkload; - void Execute() const override; - void ExecuteAsync(WorkingMemDescriptor& workingMemDescriptor) override; -private: - void Execute(std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs) const; -}; - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefWorkloads.hpp b/src/backends/reference/workloads/RefWorkloads.hpp index ed3aa90e5f..914137c23d 100644 --- a/src/backends/reference/workloads/RefWorkloads.hpp +++ b/src/backends/reference/workloads/RefWorkloads.hpp @@ -59,7 +59,6 @@ #include "RefRankWorkload.hpp" #include "RefReduceWorkload.hpp" #include "RefReshapeWorkload.hpp" -#include "RefResizeBilinearWorkload.hpp" #include "RefResizeWorkload.hpp" #include "RefShapeWorkload.hpp" #include "RefSliceWorkload.hpp" |