aboutsummaryrefslogtreecommitdiff
path: root/src/backends/reference
diff options
context:
space:
mode:
Diffstat (limited to 'src/backends/reference')
-rw-r--r--src/backends/reference/RefBackend.cpp15
-rw-r--r--src/backends/reference/RefBackend.hpp3
-rw-r--r--src/backends/reference/RefLayerSupport.cpp119
-rw-r--r--src/backends/reference/RefLayerSupport.hpp37
-rw-r--r--src/backends/reference/RefWorkloadFactory.cpp58
-rw-r--r--src/backends/reference/RefWorkloadFactory.hpp24
-rw-r--r--src/backends/reference/backend.mk1
-rw-r--r--src/backends/reference/workloads/CMakeLists.txt2
-rw-r--r--src/backends/reference/workloads/Decoders.hpp7
-rw-r--r--src/backends/reference/workloads/Encoders.hpp10
-rw-r--r--src/backends/reference/workloads/RefResizeBilinearWorkload.cpp45
-rw-r--r--src/backends/reference/workloads/RefResizeBilinearWorkload.hpp24
-rw-r--r--src/backends/reference/workloads/RefWorkloads.hpp1
13 files changed, 5 insertions, 341 deletions
diff --git a/src/backends/reference/RefBackend.cpp b/src/backends/reference/RefBackend.cpp
index c9f164e0c9..a3060f0798 100644
--- a/src/backends/reference/RefBackend.cpp
+++ b/src/backends/reference/RefBackend.cpp
@@ -58,27 +58,12 @@ IBackendInternal::IMemoryManagerUniquePtr RefBackend::CreateMemoryManager() cons
return std::make_unique<RefMemoryManager>();
}
-IBackendInternal::Optimizations RefBackend::GetOptimizations() const
-{
- return Optimizations{};
-}
-
IBackendInternal::ILayerSupportSharedPtr RefBackend::GetLayerSupport() const
{
static ILayerSupportSharedPtr layerSupport{new RefLayerSupport};
return layerSupport;
}
-bool RefBackend::HasCapability(BackendCapability capabilityClass) const
-{
- auto search = oldCpuRefCapabilities.find(capabilityClass);
- if (search != oldCpuRefCapabilities.end())
- {
- return true;
- }
- return false;
-}
-
OptimizationViews RefBackend::OptimizeSubgraphView(const SubgraphView& subgraph) const
{
OptimizationViews optimizationViews;
diff --git a/src/backends/reference/RefBackend.hpp b/src/backends/reference/RefBackend.hpp
index 28c1591179..c04bf43db3 100644
--- a/src/backends/reference/RefBackend.hpp
+++ b/src/backends/reference/RefBackend.hpp
@@ -47,7 +47,6 @@ public:
IBackendInternal::IBackendProfilingContextPtr CreateBackendProfilingContext(
const IRuntime::CreationOptions& creationOptions, IBackendProfilingPtr& backendProfiling) override;
- IBackendInternal::Optimizations GetOptimizations() const override;
IBackendInternal::ILayerSupportSharedPtr GetLayerSupport() const override;
OptimizationViews OptimizeSubgraphView(const SubgraphView& subgraph) const override;
@@ -60,8 +59,6 @@ public:
{
return cpuRefCapabilities;
};
-
- bool HasCapability(BackendCapability capabilityClass) const override;
};
} // namespace armnn
diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp
index c0ede678bf..b80aa9992f 100644
--- a/src/backends/reference/RefLayerSupport.cpp
+++ b/src/backends/reference/RefLayerSupport.cpp
@@ -58,15 +58,6 @@ std::string CreateIncorrectDimensionsErrorMsg(unsigned int expected,
} // anonymous namespace
-bool RefLayerSupport::IsAbsSupported(const TensorInfo& input, const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported) const
-{
- return IsElementwiseUnarySupported(input,
- output,
- ElementwiseUnaryDescriptor(UnaryOperation::Abs),
- reasonIfUnsupported);
-}
-
bool RefLayerSupport::IsActivationSupported(const TensorInfo& input,
const TensorInfo& output,
const ActivationDescriptor& descriptor,
@@ -565,15 +556,12 @@ bool RefLayerSupport::IsConvolution2dSupported(const TensorInfo& input,
const DataType inputType = input.GetDataType();
if (IsQuantized8BitType(inputType))
{
- ARMNN_NO_DEPRECATE_WARN_BEGIN
- std::array<DataType, 4> supportedWeightTypes =
+ std::array<DataType, 3> supportedWeightTypes =
{
DataType::QAsymmS8,
DataType::QAsymmU8,
- DataType::QSymmS8,
- DataType::QuantizedSymm8PerAxis // deprecated
+ DataType::QSymmS8
};
- ARMNN_NO_DEPRECATE_WARN_END
supported &= CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported,
"Reference Convolution2d: weights type not supported for quantized input.");
@@ -769,15 +757,12 @@ bool RefLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input,
const DataType inputType = input.GetDataType();
if (IsQuantized8BitType(inputType))
{
- ARMNN_NO_DEPRECATE_WARN_BEGIN
- std::array<DataType, 4> supportedWeightTypes =
+ std::array<DataType, 3> supportedWeightTypes =
{
DataType::QAsymmS8,
DataType::QAsymmU8,
DataType::QSymmS8,
- DataType::QuantizedSymm8PerAxis // deprecated
};
- ARMNN_NO_DEPRECATE_WARN_END
supported &= CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported,
"Reference DepthwiseConvolution2d: weights type not supported for "
@@ -977,18 +962,6 @@ bool RefLayerSupport::IsElementwiseUnarySupported(const TensorInfo& input,
return supported;
}
-bool RefLayerSupport::IsEqualSupported(const TensorInfo& input0,
- const TensorInfo& input1,
- const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported) const
-{
- return IsComparisonSupported(input0,
- input1,
- output,
- ComparisonDescriptor(ComparisonOperation::Equal),
- reasonIfUnsupported);
-}
-
bool RefLayerSupport::IsFakeQuantizationSupported(const TensorInfo& input,
const FakeQuantizationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
@@ -1173,18 +1146,6 @@ bool RefLayerSupport::IsGatherSupported(const armnn::TensorInfo& input0,
return supported;
}
-bool RefLayerSupport::IsGreaterSupported(const TensorInfo& input0,
- const TensorInfo& input1,
- const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported) const
-{
- return IsComparisonSupported(input0,
- input1,
- output,
- ComparisonDescriptor(ComparisonOperation::Greater),
- reasonIfUnsupported);
-}
-
bool RefLayerSupport::IsInputSupported(const TensorInfo& /*input*/,
Optional<std::string&> /*reasonIfUnsupported*/) const
{
@@ -1523,14 +1484,6 @@ bool RefLayerSupport::IsMeanSupported(const TensorInfo& input,
return supported;
}
-bool RefLayerSupport::IsMergerSupported(const std::vector<const TensorInfo*> inputs,
- const TensorInfo& output,
- const MergerDescriptor& descriptor,
- Optional<std::string&> reasonIfUnsupported) const
-{
- return IsConcatSupported(inputs, output, descriptor, reasonIfUnsupported);
-}
-
bool RefLayerSupport::IsMemCopySupported(const TensorInfo &input,
const TensorInfo &output,
Optional<std::string &> reasonIfUnsupported) const
@@ -1897,33 +1850,6 @@ bool RefLayerSupport::IsReshapeSupported(const TensorInfo& input,
"Reference reshape: input type not supported.");
}
-bool RefLayerSupport::IsResizeBilinearSupported(const TensorInfo& input,
- const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported) const
-{
- bool supported = true;
- std::array<DataType,6> supportedTypes =
- {
- DataType::BFloat16,
- DataType::Float32,
- DataType::Float16,
- DataType::QAsymmS8,
- DataType::QAsymmU8,
- DataType::QSymmS16
- };
-
- supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
- "Reference ResizeBilinear: input type not supported");
-
- supported &= CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported,
- "Reference ResizeBilinear: output type not supported");
-
- supported &= CheckSupportRule(TypesAreEqual(input, output), reasonIfUnsupported,
- "Reference ResizeBilinear: input and output types not matching");
-
- return supported;
-}
-
bool RefLayerSupport::IsResizeSupported(const TensorInfo& input,
const TensorInfo& output,
const ResizeDescriptor& descriptor,
@@ -1953,16 +1879,6 @@ bool RefLayerSupport::IsResizeSupported(const TensorInfo& input,
return supported;
}
-bool RefLayerSupport::IsRsqrtSupported(const TensorInfo& input,
- const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported) const
-{
- return IsElementwiseUnarySupported(input,
- output,
- ElementwiseUnaryDescriptor(UnaryOperation::Rsqrt),
- reasonIfUnsupported);
-}
-
bool RefLayerSupport::IsShapeSupported(const TensorInfo& input,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported) const
@@ -2101,28 +2017,6 @@ bool RefLayerSupport::IsSpaceToDepthSupported(const TensorInfo& input,
}
bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input,
- const ViewsDescriptor& descriptor,
- Optional<std::string&> reasonIfUnsupported) const
-{
- IgnoreUnused(descriptor);
- bool supported = true;
- std::array<DataType,6> supportedTypes =
- {
- DataType::BFloat16,
- DataType::Float32,
- DataType::Float16,
- DataType::QAsymmS8,
- DataType::QAsymmU8,
- DataType::QSymmS16
- };
-
- supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported,
- "Reference splitter: input type not supported");
-
- return supported;
-}
-
-bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input,
const std::vector<std::reference_wrapper<TensorInfo>>& outputs,
const ViewsDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
@@ -2322,15 +2216,12 @@ bool RefLayerSupport::IsTransposeConvolution2dSupported(const TensorInfo& input,
const DataType inputType = input.GetDataType();
if (IsQuantized8BitType(inputType))
{
- ARMNN_NO_DEPRECATE_WARN_BEGIN
- std::array<DataType, 4> supportedWeightTypes =
+ std::array<DataType, 3> supportedWeightTypes =
{
DataType::QAsymmS8,
DataType::QAsymmU8,
- DataType::QSymmS8,
- DataType::QuantizedSymm8PerAxis //Deprecated
+ DataType::QSymmS8
};
- ARMNN_NO_DEPRECATE_WARN_END
supported &= CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported,
"Reference TransposeConvolution2d: weights type not supported for "
diff --git a/src/backends/reference/RefLayerSupport.hpp b/src/backends/reference/RefLayerSupport.hpp
index 627418e3e1..53d7907204 100644
--- a/src/backends/reference/RefLayerSupport.hpp
+++ b/src/backends/reference/RefLayerSupport.hpp
@@ -12,11 +12,6 @@ namespace armnn
class RefLayerSupport : public LayerSupportBase
{
public:
- ARMNN_DEPRECATED_MSG("Use IsElementwiseUnarySupported instead")
- bool IsAbsSupported(const TensorInfo& input,
- const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
-
bool IsActivationSupported(const TensorInfo& input,
const TensorInfo& output,
const ActivationDescriptor& descriptor,
@@ -147,12 +142,6 @@ public:
const ElementwiseUnaryDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
- ARMNN_DEPRECATED_MSG("Use IsComparisonSupported instead")
- bool IsEqualSupported(const TensorInfo& input0,
- const TensorInfo& input1,
- const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
-
bool IsFakeQuantizationSupported(const TensorInfo& input,
const FakeQuantizationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
@@ -179,12 +168,6 @@ public:
const GatherDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
- ARMNN_DEPRECATED_MSG("Use IsComparisonSupported instead")
- bool IsGreaterSupported(const TensorInfo& input0,
- const TensorInfo& input1,
- const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
-
bool IsInputSupported(const TensorInfo& input,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
@@ -230,12 +213,6 @@ public:
const MeanDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
- ARMNN_DEPRECATED_MSG("Use IsConcatSupported instead")
- bool IsMergerSupported(const std::vector<const TensorInfo*> inputs,
- const TensorInfo& output,
- const MergerDescriptor& descriptor,
- Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
-
bool IsMemCopySupported(const TensorInfo& input,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
@@ -301,20 +278,11 @@ public:
const ReshapeDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
- bool IsResizeBilinearSupported(const TensorInfo& input,
- const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
-
bool IsResizeSupported(const TensorInfo& input,
const TensorInfo& output,
const ResizeDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
- ARMNN_DEPRECATED_MSG("Use IsElementwiseUnarySupported instead")
- bool IsRsqrtSupported(const TensorInfo& input,
- const TensorInfo& output,
- Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
-
bool IsShapeSupported(const TensorInfo& input,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
@@ -340,11 +308,6 @@ public:
Optional<std::string&> reasonIfUnsupported = EmptyOptional())
const override;
- ARMNN_DEPRECATED_MSG("Use IsSplitterSupported with outputs instead")
- bool IsSplitterSupported(const TensorInfo& input,
- const ViewsDescriptor& descriptor,
- Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
-
bool IsSplitterSupported(const TensorInfo& input,
const std::vector<std::reference_wrapper<TensorInfo>>& outputs,
const ViewsDescriptor& descriptor,
diff --git a/src/backends/reference/RefWorkloadFactory.cpp b/src/backends/reference/RefWorkloadFactory.cpp
index 18a5af277f..75008bc866 100644
--- a/src/backends/reference/RefWorkloadFactory.cpp
+++ b/src/backends/reference/RefWorkloadFactory.cpp
@@ -129,16 +129,6 @@ std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const Tens
return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager);
}
-std::unique_ptr<IWorkload> RefWorkloadFactory::CreateAbs(const AbsQueueDescriptor& descriptor,
- const WorkloadInfo& info) const
-{
- IgnoreUnused(descriptor);
- ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
- elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Abs;
-
- return CreateElementwiseUnary(elementwiseUnaryDescriptor, info);
-}
-
std::unique_ptr<IWorkload> RefWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
@@ -331,16 +321,6 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateElementwiseUnary(const Elem
return std::make_unique<RefElementwiseUnaryWorkload>(descriptor, info);
}
-std::unique_ptr<IWorkload> RefWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor,
- const WorkloadInfo& info) const
-{
- IgnoreUnused(descriptor);
- ComparisonQueueDescriptor comparisonDescriptor;
- comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Equal;
-
- return CreateComparison(comparisonDescriptor, info);
-}
-
std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
@@ -379,16 +359,6 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGather(const GatherQueueDes
return std::make_unique<RefGatherWorkload>(descriptor, info);
}
-std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor,
- const WorkloadInfo& info) const
-{
- IgnoreUnused(descriptor);
- ComparisonQueueDescriptor comparisonDescriptor;
- comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Greater;
-
- return CreateComparison(comparisonDescriptor, info);
-}
-
std::unique_ptr<IWorkload> RefWorkloadFactory::CreateInput(const InputQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
@@ -479,12 +449,6 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMemImport(const MemImportQu
return std::make_unique<ImportMemGenericWorkload>(descriptor, info);
}
-std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMerger(const MergerQueueDescriptor& descriptor,
- const WorkloadInfo& info) const
-{
- return CreateConcat(descriptor, info);
-}
-
std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
@@ -615,28 +579,6 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResize(const ResizeQueueDes
return std::make_unique<RefResizeWorkload>(descriptor, info);
}
-std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResizeBilinear(const ResizeBilinearQueueDescriptor& descriptor,
- const WorkloadInfo& info) const
-{
- ResizeQueueDescriptor resizeDescriptor;
- resizeDescriptor.m_Parameters.m_Method = ResizeMethod::Bilinear;
- resizeDescriptor.m_Parameters.m_DataLayout = descriptor.m_Parameters.m_DataLayout;
- resizeDescriptor.m_Parameters.m_TargetWidth = descriptor.m_Parameters.m_TargetWidth;
- resizeDescriptor.m_Parameters.m_TargetHeight = descriptor.m_Parameters.m_TargetHeight;
-
- return CreateResize(resizeDescriptor, info);
-}
-
-std::unique_ptr<IWorkload> RefWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& descriptor,
- const WorkloadInfo& info) const
-{
- IgnoreUnused(descriptor);
- ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
- elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Rsqrt;
-
- return CreateElementwiseUnary(elementwiseUnaryDescriptor, info);
-}
-
std::unique_ptr<IWorkload> RefWorkloadFactory::CreateShape(const ShapeQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
diff --git a/src/backends/reference/RefWorkloadFactory.hpp b/src/backends/reference/RefWorkloadFactory.hpp
index d00d3ca822..a85e8dda3e 100644
--- a/src/backends/reference/RefWorkloadFactory.hpp
+++ b/src/backends/reference/RefWorkloadFactory.hpp
@@ -66,10 +66,6 @@ public:
DataLayout dataLayout,
const bool IsMemoryManaged = true) const override;
- ARMNN_DEPRECATED_MSG("Use CreateElementwiseUnary instead")
- std::unique_ptr<IWorkload> CreateAbs(const AbsQueueDescriptor& descriptor,
- const WorkloadInfo& info) const override;
-
std::unique_ptr<IWorkload> CreateActivation(const ActivationQueueDescriptor& descriptor,
const WorkloadInfo& info) const override;
@@ -139,10 +135,6 @@ public:
std::unique_ptr<IWorkload> CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& descriptor,
const WorkloadInfo& info) const override;
- ARMNN_DEPRECATED_MSG("Use CreateComparison instead")
- std::unique_ptr<IWorkload> CreateEqual(const EqualQueueDescriptor& descriptor,
- const WorkloadInfo& info) const override;
-
std::unique_ptr<IWorkload> CreateFakeQuantization(const FakeQuantizationQueueDescriptor& descriptor,
const WorkloadInfo& info) const override;
@@ -158,10 +150,6 @@ public:
std::unique_ptr<IWorkload> CreateGather(const GatherQueueDescriptor& descriptor,
const WorkloadInfo& info) const override;
- ARMNN_DEPRECATED_MSG("Use CreateComparison instead")
- std::unique_ptr<IWorkload> CreateGreater(const GreaterQueueDescriptor& descriptor,
- const WorkloadInfo& info) const override;
-
std::unique_ptr<IWorkload> CreateInput(const InputQueueDescriptor& descriptor,
const WorkloadInfo& info) const override;
@@ -192,10 +180,6 @@ public:
std::unique_ptr<IWorkload> CreateMemImport(const MemImportQueueDescriptor& descriptor,
const WorkloadInfo& info) const override;
- ARMNN_DEPRECATED_MSG("Use CreateConcat instead")
- std::unique_ptr<IWorkload> CreateMerger(const MergerQueueDescriptor& descriptor,
- const WorkloadInfo& info) const override;
-
std::unique_ptr<IWorkload> CreateMinimum(const MinimumQueueDescriptor& descriptor,
const WorkloadInfo& info) const override;
@@ -241,14 +225,6 @@ public:
std::unique_ptr<IWorkload> CreateResize(const ResizeQueueDescriptor& descriptor,
const WorkloadInfo& info) const override;
- ARMNN_DEPRECATED_MSG("Use CreateResize instead")
- std::unique_ptr<IWorkload> CreateResizeBilinear(const ResizeBilinearQueueDescriptor& descriptor,
- const WorkloadInfo& info) const override;
-
- ARMNN_DEPRECATED_MSG("Use CreateElementwiseUnary instead")
- std::unique_ptr<IWorkload> CreateRsqrt(const RsqrtQueueDescriptor& descriptor,
- const WorkloadInfo& info) const override;
-
std::unique_ptr<IWorkload> CreateShape(const ShapeQueueDescriptor& descriptor,
const WorkloadInfo& info) const override;
diff --git a/src/backends/reference/backend.mk b/src/backends/reference/backend.mk
index 7d6c59a273..f8169a6c0c 100644
--- a/src/backends/reference/backend.mk
+++ b/src/backends/reference/backend.mk
@@ -88,7 +88,6 @@ BACKEND_SOURCES := \
workloads/RefQuantizeWorkload.cpp \
workloads/RefReduceWorkload.cpp \
workloads/RefReshapeWorkload.cpp \
- workloads/RefResizeBilinearWorkload.cpp \
workloads/RefResizeWorkload.cpp \
workloads/RefSliceWorkload.cpp \
workloads/RefSoftmaxWorkload.cpp \
diff --git a/src/backends/reference/workloads/CMakeLists.txt b/src/backends/reference/workloads/CMakeLists.txt
index e169c03ad8..5727291be3 100644
--- a/src/backends/reference/workloads/CMakeLists.txt
+++ b/src/backends/reference/workloads/CMakeLists.txt
@@ -148,8 +148,6 @@ list(APPEND armnnRefBackendWorkloads_sources
RefReduceWorkload.hpp
RefReshapeWorkload.cpp
RefReshapeWorkload.hpp
- RefResizeBilinearWorkload.cpp
- RefResizeBilinearWorkload.hpp
RefResizeWorkload.cpp
RefResizeWorkload.hpp
RefShapeWorkload.hpp
diff --git a/src/backends/reference/workloads/Decoders.hpp b/src/backends/reference/workloads/Decoders.hpp
index cd0dc5d40f..c2a456bfce 100644
--- a/src/backends/reference/workloads/Decoders.hpp
+++ b/src/backends/reference/workloads/Decoders.hpp
@@ -67,13 +67,6 @@ inline std::unique_ptr<Decoder<float>> MakeDecoder(const TensorInfo& info, const
{
switch(info.GetDataType())
{
- ARMNN_NO_DEPRECATE_WARN_BEGIN
- case armnn::DataType::QuantizedSymm8PerAxis:
- {
- std::pair<unsigned int, std::vector<float>> params = armnnUtils::GetPerAxisParams(info);
- return std::make_unique<QSymm8PerAxisDecoder>(static_cast<const int8_t*>(data), info);
- }
- ARMNN_NO_DEPRECATE_WARN_END
case DataType::QAsymmS8:
{
return std::make_unique<QASymmS8Decoder>(
diff --git a/src/backends/reference/workloads/Encoders.hpp b/src/backends/reference/workloads/Encoders.hpp
index a2d565ec4a..a7be9e172b 100644
--- a/src/backends/reference/workloads/Encoders.hpp
+++ b/src/backends/reference/workloads/Encoders.hpp
@@ -22,16 +22,6 @@ inline std::unique_ptr<Encoder<float>> MakeEncoder(const TensorInfo& info, void*
{
switch(info.GetDataType())
{
- ARMNN_NO_DEPRECATE_WARN_BEGIN
- case armnn::DataType::QuantizedSymm8PerAxis:
- {
- std::pair<unsigned int, std::vector<float>> params = armnnUtils::GetPerAxisParams(info);
- return std::make_unique<QSymm8PerAxisEncoder>(
- static_cast<int8_t*>(data),
- params.second,
- params.first);
- }
- ARMNN_NO_DEPRECATE_WARN_END
case armnn::DataType::QAsymmS8:
{
return std::make_unique<QASymmS8Encoder>(
diff --git a/src/backends/reference/workloads/RefResizeBilinearWorkload.cpp b/src/backends/reference/workloads/RefResizeBilinearWorkload.cpp
deleted file mode 100644
index 2cf5888f33..0000000000
--- a/src/backends/reference/workloads/RefResizeBilinearWorkload.cpp
+++ /dev/null
@@ -1,45 +0,0 @@
-//
-// Copyright © 2017 Arm Ltd. All rights reserved.
-// SPDX-License-Identifier: MIT
-//
-
-#include "RefResizeBilinearWorkload.hpp"
-
-#include "RefWorkloadUtils.hpp"
-#include "Resize.hpp"
-#include "BaseIterator.hpp"
-#include "Profiling.hpp"
-
-#include "BaseIterator.hpp"
-#include "Decoders.hpp"
-#include "Encoders.hpp"
-
-namespace armnn
-{
-
-void RefResizeBilinearWorkload::Execute() const
-{
- Execute(m_Data.m_Inputs, m_Data.m_Outputs);
-}
-
-void RefResizeBilinearWorkload::ExecuteAsync(WorkingMemDescriptor &workingMemDescriptor)
-{
- Execute(workingMemDescriptor.m_Inputs, workingMemDescriptor.m_Outputs);
-}
-
-void RefResizeBilinearWorkload::Execute(std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs) const
-{
- ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefResizeBilinearWorkload_Execute");
-
- const TensorInfo& inputInfo = GetTensorInfo(inputs[0]);
- const TensorInfo& outputInfo = GetTensorInfo(outputs[0]);
-
- std::unique_ptr<Decoder<float>> decoderPtr = MakeDecoder<float>(inputInfo, inputs[0]->Map());
- Decoder<float> &decoder = *decoderPtr;
- std::unique_ptr<Encoder<float>> encoderPtr = MakeEncoder<float>(outputInfo, outputs[0]->Map());
- Encoder<float> &encoder = *encoderPtr;
-
- Resize(decoder, inputInfo, encoder, outputInfo, m_Data.m_Parameters.m_DataLayout, armnn::ResizeMethod::Bilinear);
-}
-
-} //namespace armnn
diff --git a/src/backends/reference/workloads/RefResizeBilinearWorkload.hpp b/src/backends/reference/workloads/RefResizeBilinearWorkload.hpp
deleted file mode 100644
index 5ada3d1ff8..0000000000
--- a/src/backends/reference/workloads/RefResizeBilinearWorkload.hpp
+++ /dev/null
@@ -1,24 +0,0 @@
-//
-// Copyright © 2017 Arm Ltd. All rights reserved.
-// SPDX-License-Identifier: MIT
-//
-
-#pragma once
-
-#include <backendsCommon/Workload.hpp>
-#include <backendsCommon/WorkloadData.hpp>
-
-namespace armnn
-{
-
-class RefResizeBilinearWorkload : public BaseWorkload<ResizeBilinearQueueDescriptor>
-{
-public:
- using BaseWorkload<ResizeBilinearQueueDescriptor>::BaseWorkload;
- void Execute() const override;
- void ExecuteAsync(WorkingMemDescriptor& workingMemDescriptor) override;
-private:
- void Execute(std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs) const;
-};
-
-} //namespace armnn
diff --git a/src/backends/reference/workloads/RefWorkloads.hpp b/src/backends/reference/workloads/RefWorkloads.hpp
index ed3aa90e5f..914137c23d 100644
--- a/src/backends/reference/workloads/RefWorkloads.hpp
+++ b/src/backends/reference/workloads/RefWorkloads.hpp
@@ -59,7 +59,6 @@
#include "RefRankWorkload.hpp"
#include "RefReduceWorkload.hpp"
#include "RefReshapeWorkload.hpp"
-#include "RefResizeBilinearWorkload.hpp"
#include "RefResizeWorkload.hpp"
#include "RefShapeWorkload.hpp"
#include "RefSliceWorkload.hpp"