diff options
Diffstat (limited to 'src')
-rw-r--r-- | src/armnn/LayerSupport.cpp | 3 | ||||
-rw-r--r-- | src/backends/backendsCommon/LayerSupportBase.cpp | 1 | ||||
-rw-r--r-- | src/backends/backendsCommon/LayerSupportBase.hpp | 1 | ||||
-rw-r--r-- | src/backends/backendsCommon/WorkloadFactory.cpp | 5 | ||||
-rw-r--r-- | src/backends/cl/ClLayerSupport.cpp | 2 | ||||
-rw-r--r-- | src/backends/cl/ClLayerSupport.hpp | 1 | ||||
-rw-r--r-- | src/backends/neon/NeonLayerSupport.cpp | 11 | ||||
-rw-r--r-- | src/backends/neon/NeonLayerSupport.hpp | 4 | ||||
-rw-r--r-- | src/backends/neon/NeonWorkloadFactory.cpp | 2 | ||||
-rw-r--r-- | src/backends/neon/backend.mk | 1 | ||||
-rw-r--r-- | src/backends/neon/test/NeonLayerTests.cpp | 14 | ||||
-rw-r--r-- | src/backends/neon/workloads/CMakeLists.txt | 2 | ||||
-rw-r--r-- | src/backends/neon/workloads/NeonResizeBilinearWorkload.cpp | 59 | ||||
-rw-r--r-- | src/backends/neon/workloads/NeonResizeBilinearWorkload.hpp | 29 | ||||
-rw-r--r-- | src/backends/neon/workloads/NeonWorkloads.hpp | 1 | ||||
-rw-r--r-- | src/backends/reference/RefLayerSupport.cpp | 2 | ||||
-rw-r--r-- | src/backends/reference/RefLayerSupport.hpp | 1 |
17 files changed, 136 insertions, 3 deletions
diff --git a/src/armnn/LayerSupport.cpp b/src/armnn/LayerSupport.cpp index b600e4daa9..2eaf780f91 100644 --- a/src/armnn/LayerSupport.cpp +++ b/src/armnn/LayerSupport.cpp @@ -431,10 +431,11 @@ bool IsReshapeSupported(const BackendId& backend, bool IsResizeBilinearSupported(const BackendId& backend, const TensorInfo& input, + const TensorInfo& output, char* reasonIfUnsupported, size_t reasonIfUnsupportedMaxLength) { - FORWARD_LAYER_SUPPORT_FUNC(backend, IsResizeBilinearSupported, input); + FORWARD_LAYER_SUPPORT_FUNC(backend, IsResizeBilinearSupported, input, output); } bool IsRsqrtSupported(const BackendId& backend, diff --git a/src/backends/backendsCommon/LayerSupportBase.cpp b/src/backends/backendsCommon/LayerSupportBase.cpp index 9d68d35583..75790dc61a 100644 --- a/src/backends/backendsCommon/LayerSupportBase.cpp +++ b/src/backends/backendsCommon/LayerSupportBase.cpp @@ -316,6 +316,7 @@ bool LayerSupportBase::IsReshapeSupported(const TensorInfo& input, } bool LayerSupportBase::IsResizeBilinearSupported(const TensorInfo& input, + const TensorInfo& output, Optional<std::string&> reasonIfUnsupported) const { return DefaultLayerSupport(__func__, __FILE__, __LINE__, reasonIfUnsupported); diff --git a/src/backends/backendsCommon/LayerSupportBase.hpp b/src/backends/backendsCommon/LayerSupportBase.hpp index 5a71b44211..c3acdcab7f 100644 --- a/src/backends/backendsCommon/LayerSupportBase.hpp +++ b/src/backends/backendsCommon/LayerSupportBase.hpp @@ -199,6 +199,7 @@ public: Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; bool IsResizeBilinearSupported(const TensorInfo& input, + const TensorInfo& output, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; bool IsRsqrtSupported(const TensorInfo& input, diff --git a/src/backends/backendsCommon/WorkloadFactory.cpp b/src/backends/backendsCommon/WorkloadFactory.cpp index a70ec7e231..38a2402297 100644 --- a/src/backends/backendsCommon/WorkloadFactory.cpp +++ b/src/backends/backendsCommon/WorkloadFactory.cpp @@ -620,7 +620,10 @@ bool IWorkloadFactory::IsLayerSupported(const BackendId& backendId, case LayerType::ResizeBilinear: { const TensorInfo& input = layer.GetInputSlot(0).GetConnection()->GetTensorInfo(); - result = layerSupportObject->IsResizeBilinearSupported(OverrideDataType(input, dataType), reason); + const TensorInfo& output = layer.GetOutputSlot(0).GetTensorInfo(); + result = layerSupportObject->IsResizeBilinearSupported(OverrideDataType(input, dataType), + OverrideDataType(output, dataType), + reason); break; } case LayerType::Rsqrt: diff --git a/src/backends/cl/ClLayerSupport.cpp b/src/backends/cl/ClLayerSupport.cpp index c1139e2e1e..cb03e8b5ae 100644 --- a/src/backends/cl/ClLayerSupport.cpp +++ b/src/backends/cl/ClLayerSupport.cpp @@ -498,8 +498,10 @@ bool ClLayerSupport::IsReshapeSupported(const TensorInfo& input, } bool ClLayerSupport::IsResizeBilinearSupported(const TensorInfo& input, + const TensorInfo& output, Optional<std::string&> reasonIfUnsupported) const { + ignore_unused(output); return IsSupportedForDataTypeCl(reasonIfUnsupported, input.GetDataType(), &TrueFunc<>, diff --git a/src/backends/cl/ClLayerSupport.hpp b/src/backends/cl/ClLayerSupport.hpp index b06e6a9404..07e24335dc 100644 --- a/src/backends/cl/ClLayerSupport.hpp +++ b/src/backends/cl/ClLayerSupport.hpp @@ -169,6 +169,7 @@ public: Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; bool IsResizeBilinearSupported(const TensorInfo& input, + const TensorInfo& output, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; bool IsSoftmaxSupported(const TensorInfo& input, diff --git a/src/backends/neon/NeonLayerSupport.cpp b/src/backends/neon/NeonLayerSupport.cpp index 1f205ed6d6..76cdf140d2 100644 --- a/src/backends/neon/NeonLayerSupport.cpp +++ b/src/backends/neon/NeonLayerSupport.cpp @@ -32,6 +32,7 @@ #include "workloads/NeonFullyConnectedWorkload.hpp" #include "workloads/NeonPermuteWorkload.hpp" #include "workloads/NeonPooling2dWorkload.hpp" +#include "workloads/NeonResizeBilinearWorkload.hpp" #include "workloads/NeonSoftmaxBaseWorkload.hpp" #include "workloads/NeonSubtractionFloatWorkload.hpp" #endif @@ -374,6 +375,16 @@ bool NeonLayerSupport::IsReshapeSupported(const TensorInfo& input, &TrueFunc<>); } +bool NeonLayerSupport::IsResizeBilinearSupported(const TensorInfo& input, + const TensorInfo& output, + Optional<std::string&> reasonIfUnsupported) const +{ + FORWARD_WORKLOAD_VALIDATE_FUNC(NeonResizeBilinearWorkloadValidate, + reasonIfUnsupported, + input, + output); +} + bool NeonLayerSupport::IsSoftmaxSupported(const TensorInfo& input, const TensorInfo& output, const SoftmaxDescriptor& descriptor, diff --git a/src/backends/neon/NeonLayerSupport.hpp b/src/backends/neon/NeonLayerSupport.hpp index c522c6ee0a..631632485c 100644 --- a/src/backends/neon/NeonLayerSupport.hpp +++ b/src/backends/neon/NeonLayerSupport.hpp @@ -126,6 +126,10 @@ public: const ReshapeDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; + bool IsResizeBilinearSupported(const TensorInfo& input, + const TensorInfo& output, + Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; + bool IsSoftmaxSupported(const TensorInfo& input, const TensorInfo& output, const SoftmaxDescriptor& descriptor, diff --git a/src/backends/neon/NeonWorkloadFactory.cpp b/src/backends/neon/NeonWorkloadFactory.cpp index 311479a5f7..e7fac97c2c 100644 --- a/src/backends/neon/NeonWorkloadFactory.cpp +++ b/src/backends/neon/NeonWorkloadFactory.cpp @@ -203,7 +203,7 @@ std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateResizeBilinear( const ResizeBilinearQueueDescriptor& descriptor, const WorkloadInfo& info) const { - return nullptr; + return std::make_unique<NeonResizeBilinearWorkload>(descriptor, info); } std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateFakeQuantization( diff --git a/src/backends/neon/backend.mk b/src/backends/neon/backend.mk index 2e3848c3a3..1ebeae54f9 100644 --- a/src/backends/neon/backend.mk +++ b/src/backends/neon/backend.mk @@ -34,6 +34,7 @@ BACKEND_SOURCES := \ workloads/NeonPermuteWorkload.cpp \ workloads/NeonPooling2dWorkload.cpp \ workloads/NeonReshapeWorkload.cpp \ + workloads/NeonResizeBilinearWorkload.cpp \ workloads/NeonSoftmaxBaseWorkload.cpp \ workloads/NeonSoftmaxFloatWorkload.cpp \ workloads/NeonSoftmaxUint8Workload.cpp \ diff --git a/src/backends/neon/test/NeonLayerTests.cpp b/src/backends/neon/test/NeonLayerTests.cpp index 3d34934d06..6975374bdd 100644 --- a/src/backends/neon/test/NeonLayerTests.cpp +++ b/src/backends/neon/test/NeonLayerTests.cpp @@ -434,6 +434,20 @@ ARMNN_AUTO_TEST_CASE(SimpleNormalizationAcross, SimpleNormalizationAcrossTest) ARMNN_AUTO_TEST_CASE(SimpleNormalizationWithin, SimpleNormalizationWithinTest) ARMNN_AUTO_TEST_CASE(SimpleNormalizationAcrossNhwc, SimpleNormalizationAcrossNhwcTest) +// Resize Bilinear - NCHW data layout +ARMNN_AUTO_TEST_CASE(ResizeBilinearNop, ResizeBilinearNopTest, armnn::DataLayout::NCHW) +ARMNN_AUTO_TEST_CASE(SimpleResizeBilinear, SimpleResizeBilinearTest, armnn::DataLayout::NCHW) +ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMin, ResizeBilinearSqMinTest, armnn::DataLayout::NCHW) +ARMNN_AUTO_TEST_CASE(ResizeBilinearMin, ResizeBilinearMinTest, armnn::DataLayout::NCHW) +ARMNN_AUTO_TEST_CASE(ResizeBilinearMag, ResizeBilinearMagTest, armnn::DataLayout::NCHW) + +// Resize Bilinear - NHWC data layout +ARMNN_AUTO_TEST_CASE(ResizeBilinearNopNhwc, ResizeBilinearNopTest, armnn::DataLayout::NHWC) +ARMNN_AUTO_TEST_CASE(SimpleResizeBilinearNhwc, SimpleResizeBilinearTest, armnn::DataLayout::NHWC) +ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMinNhwc, ResizeBilinearSqMinTest, armnn::DataLayout::NHWC) +ARMNN_AUTO_TEST_CASE(ResizeBilinearMinNhwc, ResizeBilinearMinTest, armnn::DataLayout::NHWC) +ARMNN_AUTO_TEST_CASE(ResizeBilinearMagNhwc, ResizeBilinearMagTest, armnn::DataLayout::NHWC) + // ============================================================================ // COMPARE tests diff --git a/src/backends/neon/workloads/CMakeLists.txt b/src/backends/neon/workloads/CMakeLists.txt index 919c716860..713418da30 100644 --- a/src/backends/neon/workloads/CMakeLists.txt +++ b/src/backends/neon/workloads/CMakeLists.txt @@ -46,6 +46,8 @@ list(APPEND armnnNeonBackendWorkloads_sources NeonPooling2dWorkload.hpp NeonReshapeWorkload.cpp NeonReshapeWorkload.hpp + NeonResizeBilinearWorkload.cpp + NeonResizeBilinearWorkload.hpp NeonSoftmaxBaseWorkload.cpp NeonSoftmaxBaseWorkload.hpp NeonSoftmaxFloatWorkload.cpp diff --git a/src/backends/neon/workloads/NeonResizeBilinearWorkload.cpp b/src/backends/neon/workloads/NeonResizeBilinearWorkload.cpp new file mode 100644 index 0000000000..37f97bfe0e --- /dev/null +++ b/src/backends/neon/workloads/NeonResizeBilinearWorkload.cpp @@ -0,0 +1,59 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#include "NeonResizeBilinearWorkload.hpp" + +#include <aclCommon/ArmComputeUtils.hpp> +#include <aclCommon/ArmComputeTensorUtils.hpp> +#include <backendsCommon/CpuTensorHandle.hpp> +#include <neon/NeonTensorHandle.hpp> +#include <neon/NeonLayerSupport.hpp> + +using namespace armnn::armcomputetensorutils; + +namespace armnn +{ + +arm_compute::Status NeonResizeBilinearWorkloadValidate(const TensorInfo& input, const TensorInfo& output) +{ + const arm_compute::TensorInfo aclInputInfo = armcomputetensorutils::BuildArmComputeTensorInfo(input); + const arm_compute::TensorInfo aclOutputInfo = armcomputetensorutils::BuildArmComputeTensorInfo(output); + + return arm_compute::NEScale::validate(&aclInputInfo, + &aclOutputInfo, + arm_compute::InterpolationPolicy::BILINEAR, + arm_compute::BorderMode::REPLICATE, + arm_compute::PixelValue(0.f), + arm_compute::SamplingPolicy::TOP_LEFT); +} + +NeonResizeBilinearWorkload::NeonResizeBilinearWorkload(const ResizeBilinearQueueDescriptor& descriptor, + const WorkloadInfo& info) + : BaseWorkload<ResizeBilinearQueueDescriptor>(descriptor, info) +{ + m_Data.ValidateInputsOutputs("NeonResizeBilinearWorkload", 1, 1); + + arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor(); + arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor(); + + arm_compute::DataLayout aclDataLayout = ConvertDataLayout(m_Data.m_Parameters.m_DataLayout); + input.info()->set_data_layout(aclDataLayout); + output.info()->set_data_layout(aclDataLayout); + + m_ResizeBilinearLayer.configure(&input, + &output, + arm_compute::InterpolationPolicy::BILINEAR, + arm_compute::BorderMode::REPLICATE, + arm_compute::PixelValue(0.f), + arm_compute::SamplingPolicy::TOP_LEFT); +}; + +void NeonResizeBilinearWorkload::Execute() const +{ + ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonResizeBilinearWorkload_Execute"); + m_ResizeBilinearLayer.run(); +} + +} //namespace armnn diff --git a/src/backends/neon/workloads/NeonResizeBilinearWorkload.hpp b/src/backends/neon/workloads/NeonResizeBilinearWorkload.hpp new file mode 100644 index 0000000000..21753b3127 --- /dev/null +++ b/src/backends/neon/workloads/NeonResizeBilinearWorkload.hpp @@ -0,0 +1,29 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#pragma once + +#include <backendsCommon/Workload.hpp> + +#include <neon/workloads/NeonWorkloadUtils.hpp> + +#include <arm_compute/runtime/NEON/functions/NEScale.h> + +namespace armnn +{ + +arm_compute::Status NeonResizeBilinearWorkloadValidate(const TensorInfo& input, const TensorInfo& output); + +class NeonResizeBilinearWorkload : public BaseWorkload<ResizeBilinearQueueDescriptor> +{ +public: + NeonResizeBilinearWorkload(const ResizeBilinearQueueDescriptor& descriptor, const WorkloadInfo& info); + void Execute() const override; + +private: + mutable arm_compute::NEScale m_ResizeBilinearLayer; +}; + +} //namespace armnn diff --git a/src/backends/neon/workloads/NeonWorkloads.hpp b/src/backends/neon/workloads/NeonWorkloads.hpp index 70f9e376d5..e034cc94dd 100644 --- a/src/backends/neon/workloads/NeonWorkloads.hpp +++ b/src/backends/neon/workloads/NeonWorkloads.hpp @@ -25,6 +25,7 @@ #include "NeonPermuteWorkload.hpp" #include "NeonPooling2dWorkload.hpp" #include "NeonReshapeWorkload.hpp" +#include "NeonResizeBilinearWorkload.hpp" #include "NeonSoftmaxFloatWorkload.hpp" #include "NeonSoftmaxUint8Workload.hpp" #include "NeonSplitterWorkload.hpp" diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp index ce81f8d38a..25c2bafe2f 100644 --- a/src/backends/reference/RefLayerSupport.cpp +++ b/src/backends/reference/RefLayerSupport.cpp @@ -513,8 +513,10 @@ bool RefLayerSupport::IsReshapeSupported(const TensorInfo& input, } bool RefLayerSupport::IsResizeBilinearSupported(const TensorInfo& input, + const TensorInfo& output, Optional<std::string&> reasonIfUnsupported) const { + ignore_unused(output); return IsSupportedForDataTypeRef(reasonIfUnsupported, input.GetDataType(), &TrueFunc<>, diff --git a/src/backends/reference/RefLayerSupport.hpp b/src/backends/reference/RefLayerSupport.hpp index 01abc73dd5..5fe111ba15 100644 --- a/src/backends/reference/RefLayerSupport.hpp +++ b/src/backends/reference/RefLayerSupport.hpp @@ -193,6 +193,7 @@ public: Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; bool IsResizeBilinearSupported(const TensorInfo& input, + const TensorInfo& output, Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override; bool IsRsqrtSupported(const TensorInfo& input, |