From e61f0712dc480f70372fcfe0921a85a7de53661b Mon Sep 17 00:00:00 2001 From: Pablo Tello Date: Thu, 23 Jan 2020 10:37:17 +0000 Subject: MLCE-156: Add Division workload in CpuAcc Change-Id: I1f228fcaf1077867d9755a2b850c6703387fab34 Signed-off-by: Pablo Tello --- src/backends/neon/NeonLayerSupport.cpp | 13 +++++++ src/backends/neon/NeonLayerSupport.hpp | 5 +++ src/backends/neon/NeonWorkloadFactory.cpp | 4 +- src/backends/neon/backend.mk | 1 + src/backends/neon/test/NeonCreateWorkloadTests.cpp | 8 ++++ src/backends/neon/test/NeonLayerTests.cpp | 6 +++ src/backends/neon/workloads/CMakeLists.txt | 2 + .../neon/workloads/NeonDivisionWorkload.cpp | 45 ++++++++++++++++++++++ .../neon/workloads/NeonDivisionWorkload.hpp | 30 +++++++++++++++ src/backends/neon/workloads/NeonWorkloads.hpp | 1 + 10 files changed, 113 insertions(+), 2 deletions(-) create mode 100644 src/backends/neon/workloads/NeonDivisionWorkload.cpp create mode 100644 src/backends/neon/workloads/NeonDivisionWorkload.hpp diff --git a/src/backends/neon/NeonLayerSupport.cpp b/src/backends/neon/NeonLayerSupport.cpp index 47bcc2e795..6ca69f4841 100644 --- a/src/backends/neon/NeonLayerSupport.cpp +++ b/src/backends/neon/NeonLayerSupport.cpp @@ -38,6 +38,7 @@ #include "workloads/NeonConcatWorkload.hpp" #include "workloads/NeonMinimumWorkload.hpp" #include "workloads/NeonMultiplicationWorkload.hpp" +#include "workloads/NeonDivisionWorkload.hpp" #include "workloads/NeonNormalizationFloatWorkload.hpp" #include "workloads/NeonFullyConnectedWorkload.hpp" #include "workloads/NeonPadWorkload.hpp" @@ -554,6 +555,18 @@ bool NeonLayerSupport::IsMultiplicationSupported(const TensorInfo& input0, output); } +bool NeonLayerSupport::IsDivisionSupported(const TensorInfo& input0, + const TensorInfo& input1, + const TensorInfo& output, + Optional reasonIfUnsupported) const +{ + FORWARD_WORKLOAD_VALIDATE_FUNC(NeonDivisionWorkloadValidate, + reasonIfUnsupported, + input0, + input1, + output); +} + bool NeonLayerSupport::IsNormalizationSupported(const TensorInfo& input, const TensorInfo& output, const NormalizationDescriptor& descriptor, diff --git a/src/backends/neon/NeonLayerSupport.hpp b/src/backends/neon/NeonLayerSupport.hpp index e49e78b3d2..f1d87f65f3 100644 --- a/src/backends/neon/NeonLayerSupport.hpp +++ b/src/backends/neon/NeonLayerSupport.hpp @@ -176,6 +176,11 @@ public: const TensorInfo& output, Optional reasonIfUnsupported = EmptyOptional()) const override; + bool IsDivisionSupported(const TensorInfo& input0, + const TensorInfo& input1, + const TensorInfo& output, + Optional reasonIfUnsupported = EmptyOptional()) const override; + bool IsNormalizationSupported(const TensorInfo& input, const TensorInfo& output, const NormalizationDescriptor& descriptor, diff --git a/src/backends/neon/NeonWorkloadFactory.cpp b/src/backends/neon/NeonWorkloadFactory.cpp index 649cb9f699..cb2e88eddb 100644 --- a/src/backends/neon/NeonWorkloadFactory.cpp +++ b/src/backends/neon/NeonWorkloadFactory.cpp @@ -216,10 +216,10 @@ std::unique_ptr NeonWorkloadFactory::CreateDetectionPostProcess( std::unique_ptr NeonWorkloadFactory::CreateDivision( const DivisionQueueDescriptor& descriptor, const WorkloadInfo& info) const { - return MakeWorkloadHelper(descriptor, info); + return std::make_unique(descriptor, info); } -std::unique_ptr NeonWorkloadFactory::CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& +std::unique_ptr NeonWorkloadFactory::CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor& descriptor, const WorkloadInfo& info) const { diff --git a/src/backends/neon/backend.mk b/src/backends/neon/backend.mk index 05dcd02c9a..740cbcd1ff 100644 --- a/src/backends/neon/backend.mk +++ b/src/backends/neon/backend.mk @@ -45,6 +45,7 @@ BACKEND_SOURCES := \ workloads/NeonMeanWorkload.cpp \ workloads/NeonMinimumWorkload.cpp \ workloads/NeonMultiplicationWorkload.cpp \ + workloads/NeonDivisionWorkload.cpp \ workloads/NeonNormalizationFloatWorkload.cpp \ workloads/NeonPadWorkload.cpp \ workloads/NeonPermuteWorkload.cpp \ diff --git a/src/backends/neon/test/NeonCreateWorkloadTests.cpp b/src/backends/neon/test/NeonCreateWorkloadTests.cpp index 400a5a38e2..3e1888cb54 100644 --- a/src/backends/neon/test/NeonCreateWorkloadTests.cpp +++ b/src/backends/neon/test/NeonCreateWorkloadTests.cpp @@ -181,6 +181,14 @@ BOOST_AUTO_TEST_CASE(CreateMultiplicationUint8Workload) DataType::QAsymmU8>(); } +BOOST_AUTO_TEST_CASE(CreateDivisionFloatWorkloadTest) +{ + NeonCreateElementwiseWorkloadTest(); +} + template static void NeonCreateBatchNormalizationWorkloadTest(DataLayout dataLayout) { diff --git a/src/backends/neon/test/NeonLayerTests.cpp b/src/backends/neon/test/NeonLayerTests.cpp index f14b2a446b..1b25cad993 100644 --- a/src/backends/neon/test/NeonLayerTests.cpp +++ b/src/backends/neon/test/NeonLayerTests.cpp @@ -491,6 +491,12 @@ ARMNN_AUTO_TEST_CASE(SubtractionUint8, SubtractionUint8Test) ARMNN_AUTO_TEST_CASE(SubBroadcastUint8, SubtractionBroadcastUint8Test) ARMNN_AUTO_TEST_CASE(SubBroadcast1ElementUint8, SubtractionBroadcast1ElementUint8Test) +// Div +ARMNN_AUTO_TEST_CASE(SimpleDivision, DivisionTest) +ARMNN_AUTO_TEST_CASE(DivisionByZero, DivisionByZeroTest) +ARMNN_AUTO_TEST_CASE(DivisionBroadcast1Element, DivisionBroadcast1ElementTest) +ARMNN_AUTO_TEST_CASE(DivisionBroadcast1DVector, DivisionBroadcast1DVectorTest) + // Mul ARMNN_AUTO_TEST_CASE(SimpleMultiplication, MultiplicationTest) ARMNN_AUTO_TEST_CASE(MultiplicationBroadcast1Element, MultiplicationBroadcast1ElementTest) diff --git a/src/backends/neon/workloads/CMakeLists.txt b/src/backends/neon/workloads/CMakeLists.txt index d2c549ee25..46b5332235 100644 --- a/src/backends/neon/workloads/CMakeLists.txt +++ b/src/backends/neon/workloads/CMakeLists.txt @@ -52,6 +52,8 @@ list(APPEND armnnNeonBackendWorkloads_sources NeonMinimumWorkload.hpp NeonMultiplicationWorkload.cpp NeonMultiplicationWorkload.hpp + NeonDivisionWorkload.cpp + NeonDivisionWorkload.hpp NeonNormalizationFloatWorkload.cpp NeonNormalizationFloatWorkload.hpp NeonPadWorkload.cpp diff --git a/src/backends/neon/workloads/NeonDivisionWorkload.cpp b/src/backends/neon/workloads/NeonDivisionWorkload.cpp new file mode 100644 index 0000000000..6fdb455f25 --- /dev/null +++ b/src/backends/neon/workloads/NeonDivisionWorkload.cpp @@ -0,0 +1,45 @@ +// +// Copyright © 2020 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#include "NeonDivisionWorkload.hpp" +#include +#include + +namespace armnn +{ + +arm_compute::Status NeonDivisionWorkloadValidate(const TensorInfo& input0, + const TensorInfo& input1, + const TensorInfo& output) +{ + const arm_compute::TensorInfo aclInput0 = armcomputetensorutils::BuildArmComputeTensorInfo(input0); + const arm_compute::TensorInfo aclInput1 = armcomputetensorutils::BuildArmComputeTensorInfo(input1); + const arm_compute::TensorInfo aclOutput = armcomputetensorutils::BuildArmComputeTensorInfo(output); + + return arm_compute::NEElementwiseDivision::validate(&aclInput0, + &aclInput1, + &aclOutput); +} + +NeonDivisionWorkload::NeonDivisionWorkload(const DivisionQueueDescriptor& descriptor, + const WorkloadInfo& info) + : BaseWorkload(descriptor, info) +{ + m_Data.ValidateInputsOutputs("NeonDivisionWorkload", 2, 1); + + arm_compute::ITensor& input0 = boost::polymorphic_downcast(m_Data.m_Inputs[0])->GetTensor(); + arm_compute::ITensor& input1 = boost::polymorphic_downcast(m_Data.m_Inputs[1])->GetTensor(); + arm_compute::ITensor& output = boost::polymorphic_downcast(m_Data.m_Outputs[0])->GetTensor(); + + m_DivLayer.configure(&input0, &input1, &output); +} + +void NeonDivisionWorkload::Execute() const +{ + ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonDivisionWorkload_Execute"); + m_DivLayer.run(); +} + +} //namespace armnn diff --git a/src/backends/neon/workloads/NeonDivisionWorkload.hpp b/src/backends/neon/workloads/NeonDivisionWorkload.hpp new file mode 100644 index 0000000000..2405d9a4ab --- /dev/null +++ b/src/backends/neon/workloads/NeonDivisionWorkload.hpp @@ -0,0 +1,30 @@ +// +// Copyright © 2020 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#pragma once + +#include + +#include + +namespace armnn +{ + +arm_compute::Status NeonDivisionWorkloadValidate(const TensorInfo& input0, + const TensorInfo& input1, + const TensorInfo& output); + +class NeonDivisionWorkload : public BaseWorkload +{ +public: + NeonDivisionWorkload(const DivisionQueueDescriptor& descriptor, const WorkloadInfo& info); + + virtual void Execute() const override; + +private: + mutable arm_compute::NEElementwiseDivision m_DivLayer; +}; + +} //namespace armnn diff --git a/src/backends/neon/workloads/NeonWorkloads.hpp b/src/backends/neon/workloads/NeonWorkloads.hpp index 104504e097..39cf044b6c 100644 --- a/src/backends/neon/workloads/NeonWorkloads.hpp +++ b/src/backends/neon/workloads/NeonWorkloads.hpp @@ -7,6 +7,7 @@ #include "NeonAbsWorkload.hpp" #include "NeonActivationWorkload.hpp" #include "NeonAdditionWorkload.hpp" +#include "NeonDivisionWorkload.hpp" #include "NeonArgMinMaxWorkload.hpp" #include "NeonBatchNormalizationWorkload.hpp" #include "NeonConstantWorkload.hpp" -- cgit v1.2.1