From a121de3edee75180d335d8be42cdd6620a6063a5 Mon Sep 17 00:00:00 2001 From: nikraj01 Date: Wed, 29 May 2019 10:51:05 +0100 Subject: IVGCVSW-3168 Refactor reference softmax workload into a single workload Change-Id: Ie290efcbb9e3a6365cbd630cb2041e7b0f542505 Signed-off-by: nikraj01 --- .../backendsCommon/test/JsonPrinterTestImpl.cpp | 5 ++- .../backendsCommon/test/WorkloadDataValidation.cpp | 2 +- src/backends/reference/RefWorkloadFactory.cpp | 6 +++- src/backends/reference/backend.mk | 3 +- .../reference/test/RefCreateWorkloadTests.cpp | 6 ++-- src/backends/reference/workloads/CMakeLists.txt | 6 ++-- .../workloads/RefSoftmaxFloat32Workload.cpp | 26 --------------- .../workloads/RefSoftmaxFloat32Workload.hpp | 21 ------------ .../workloads/RefSoftmaxUint8Workload.cpp | 36 -------------------- .../workloads/RefSoftmaxUint8Workload.hpp | 21 ------------ .../reference/workloads/RefSoftmaxWorkload.cpp | 39 ++++++++++++++++++++++ .../reference/workloads/RefSoftmaxWorkload.hpp | 21 ++++++++++++ src/backends/reference/workloads/RefWorkloads.hpp | 3 +- src/backends/reference/workloads/Softmax.cpp | 19 +++++++---- src/backends/reference/workloads/Softmax.hpp | 3 +- 15 files changed, 91 insertions(+), 126 deletions(-) delete mode 100644 src/backends/reference/workloads/RefSoftmaxFloat32Workload.cpp delete mode 100644 src/backends/reference/workloads/RefSoftmaxFloat32Workload.hpp delete mode 100644 src/backends/reference/workloads/RefSoftmaxUint8Workload.cpp delete mode 100644 src/backends/reference/workloads/RefSoftmaxUint8Workload.hpp create mode 100644 src/backends/reference/workloads/RefSoftmaxWorkload.cpp create mode 100644 src/backends/reference/workloads/RefSoftmaxWorkload.hpp diff --git a/src/backends/backendsCommon/test/JsonPrinterTestImpl.cpp b/src/backends/backendsCommon/test/JsonPrinterTestImpl.cpp index 3248954e3b..4591def34f 100644 --- a/src/backends/backendsCommon/test/JsonPrinterTestImpl.cpp +++ b/src/backends/backendsCommon/test/JsonPrinterTestImpl.cpp @@ -244,6 +244,7 @@ void RunSoftmaxProfilerJsonPrinterTest(const std::vector& back std::string result = GetSoftmaxProfilerJson(backends); std::string backend = "Ref"; + std::string testName = "SoftmaxWorkload_Execute"; std::string changeLine31 = "\n},\n\"CopyMemGeneric_Execute\": {"; std::string changeLine39 = "us\""; std::string changeLine40; @@ -253,6 +254,7 @@ void RunSoftmaxProfilerJsonPrinterTest(const std::vector& back if (firstBackend == armnn::Compute::GpuAcc) { backend = "Cl"; + testName = "SoftmaxUintWorkload_Execute"; changeLine31 = ",\n\"OpenClKernelTimer/: softmax_layer_max_shift_exp_sum_quantized_serial GWS[,,]\": {"; changeLine39 = R"(us" }, @@ -278,6 +280,7 @@ void RunSoftmaxProfilerJsonPrinterTest(const std::vector& back else if (firstBackend == armnn::Compute::CpuAcc) { backend = "Neon"; + testName = "SoftmaxUintWorkload_Execute"; changeLine31 = ",\n\"NeonKernelTimer/: NEFillBorderKernel\": {"; changeLine39 = R"(us" }, @@ -332,7 +335,7 @@ void RunSoftmaxProfilerJsonPrinterTest(const std::vector& back ], "unit": "us" }, -")" + backend + R"(SoftmaxUintWorkload_Execute": { +")" + backend + testName + R"(": { "raw": [ , , diff --git a/src/backends/backendsCommon/test/WorkloadDataValidation.cpp b/src/backends/backendsCommon/test/WorkloadDataValidation.cpp index 94bef9b50a..795791fc5e 100644 --- a/src/backends/backendsCommon/test/WorkloadDataValidation.cpp +++ b/src/backends/backendsCommon/test/WorkloadDataValidation.cpp @@ -78,7 +78,7 @@ BOOST_AUTO_TEST_CASE(SoftmaxQueueDescriptor_Validate_WrongInputHeight) AddOutputToWorkload(invalidData, invalidInfo, outputTensorInfo, nullptr); //Invalid argument exception is expected, because height != 1. - BOOST_CHECK_THROW(RefSoftmaxFloat32Workload(invalidData, invalidInfo), armnn::InvalidArgumentException); + BOOST_CHECK_THROW(RefSoftmaxWorkload(invalidData, invalidInfo), armnn::InvalidArgumentException); } BOOST_AUTO_TEST_CASE(FullyConnectedQueueDescriptor_Validate_RequiredDataMissing) diff --git a/src/backends/reference/RefWorkloadFactory.cpp b/src/backends/reference/RefWorkloadFactory.cpp index a21becdb13..50e3c0006c 100644 --- a/src/backends/reference/RefWorkloadFactory.cpp +++ b/src/backends/reference/RefWorkloadFactory.cpp @@ -123,7 +123,11 @@ std::unique_ptr RefWorkloadFactory::CreateActivation(const Activation std::unique_ptr RefWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor, const WorkloadInfo& info) const { - return MakeWorkload(descriptor, info); + if (IsFloat16(info)) + { + return MakeWorkload(descriptor, info); + } + return std::make_unique(descriptor, info); } std::unique_ptr RefWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor, diff --git a/src/backends/reference/backend.mk b/src/backends/reference/backend.mk index 9a4cf146c6..57204a05ac 100644 --- a/src/backends/reference/backend.mk +++ b/src/backends/reference/backend.mk @@ -58,8 +58,7 @@ BACKEND_SOURCES := \ workloads/RefResizeBilinearFloat32Workload.cpp \ workloads/RefResizeBilinearUint8Workload.cpp \ workloads/RefRsqrtFloat32Workload.cpp \ - workloads/RefSoftmaxFloat32Workload.cpp \ - workloads/RefSoftmaxUint8Workload.cpp \ + workloads/RefSoftmaxWorkload.cpp \ workloads/RefSpaceToBatchNdWorkload.cpp \ workloads/RefStridedSliceWorkload.cpp \ workloads/RefSplitterFloat32Workload.cpp \ diff --git a/src/backends/reference/test/RefCreateWorkloadTests.cpp b/src/backends/reference/test/RefCreateWorkloadTests.cpp index a96d656d9b..2222a22cb3 100644 --- a/src/backends/reference/test/RefCreateWorkloadTests.cpp +++ b/src/backends/reference/test/RefCreateWorkloadTests.cpp @@ -433,12 +433,12 @@ static void RefCreateSoftmaxWorkloadTest() BOOST_AUTO_TEST_CASE(CreateSoftmaxFloat32Workload) { - RefCreateSoftmaxWorkloadTest(); + RefCreateSoftmaxWorkloadTest(); } -BOOST_AUTO_TEST_CASE(CreateSoftmaxUint8Workload) +BOOST_AUTO_TEST_CASE(CreateSoftmaxQuantisedAsymm8Workload) { - RefCreateSoftmaxWorkloadTest(); + RefCreateSoftmaxWorkloadTest(); } template diff --git a/src/backends/reference/workloads/CMakeLists.txt b/src/backends/reference/workloads/CMakeLists.txt index 3db0314346..e2f93d72a9 100644 --- a/src/backends/reference/workloads/CMakeLists.txt +++ b/src/backends/reference/workloads/CMakeLists.txt @@ -99,10 +99,8 @@ list(APPEND armnnRefBackendWorkloads_sources RefResizeBilinearUint8Workload.hpp RefRsqrtFloat32Workload.cpp RefRsqrtFloat32Workload.hpp - RefSoftmaxFloat32Workload.cpp - RefSoftmaxFloat32Workload.hpp - RefSoftmaxUint8Workload.cpp - RefSoftmaxUint8Workload.hpp + RefSoftmaxWorkload.cpp + RefSoftmaxWorkload.hpp RefSpaceToBatchNdWorkload.cpp RefSpaceToBatchNdWorkload.hpp RefSplitterFloat32Workload.cpp diff --git a/src/backends/reference/workloads/RefSoftmaxFloat32Workload.cpp b/src/backends/reference/workloads/RefSoftmaxFloat32Workload.cpp deleted file mode 100644 index 1f519bda10..0000000000 --- a/src/backends/reference/workloads/RefSoftmaxFloat32Workload.cpp +++ /dev/null @@ -1,26 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#include "RefSoftmaxFloat32Workload.hpp" - -#include "RefWorkloadUtils.hpp" -#include "Softmax.hpp" - -#include "Profiling.hpp" - -namespace armnn -{ - -void RefSoftmaxFloat32Workload::Execute() const -{ - ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefSoftmaxFloat32Workload_Execute"); - - Softmax(GetInputTensorDataFloat(0, m_Data), - GetOutputTensorDataFloat(0, m_Data), - GetTensorInfo(m_Data.m_Inputs[0]), - m_Data.m_Parameters.m_Beta); -} - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefSoftmaxFloat32Workload.hpp b/src/backends/reference/workloads/RefSoftmaxFloat32Workload.hpp deleted file mode 100644 index 82ddfac303..0000000000 --- a/src/backends/reference/workloads/RefSoftmaxFloat32Workload.hpp +++ /dev/null @@ -1,21 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#pragma once - -#include -#include - -namespace armnn -{ - -class RefSoftmaxFloat32Workload : public Float32Workload -{ -public: - using Float32Workload::Float32Workload; - virtual void Execute() const override; -}; - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefSoftmaxUint8Workload.cpp b/src/backends/reference/workloads/RefSoftmaxUint8Workload.cpp deleted file mode 100644 index 17114ec83a..0000000000 --- a/src/backends/reference/workloads/RefSoftmaxUint8Workload.cpp +++ /dev/null @@ -1,36 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#include "RefSoftmaxUint8Workload.hpp" - -#include "RefWorkloadUtils.hpp" -#include "Softmax.hpp" - -#include "Profiling.hpp" - -#include - -namespace armnn -{ - -void RefSoftmaxUint8Workload::Execute() const -{ - ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefSoftmaxUint8Workload_Execute"); - - const TensorInfo& tensorInfo = GetTensorInfo(m_Data.m_Inputs[0]); - - auto dequant = Dequantize(GetInputTensorDataU8(0, m_Data), tensorInfo); - - std::vector results(tensorInfo.GetNumElements()); - - Softmax(dequant.data(), - results.data(), - tensorInfo, - m_Data.m_Parameters.m_Beta); - - Quantize(GetOutputTensorDataU8(0, m_Data), results.data(), GetTensorInfo(m_Data.m_Outputs[0])); -} - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefSoftmaxUint8Workload.hpp b/src/backends/reference/workloads/RefSoftmaxUint8Workload.hpp deleted file mode 100644 index bb7b2143c1..0000000000 --- a/src/backends/reference/workloads/RefSoftmaxUint8Workload.hpp +++ /dev/null @@ -1,21 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#pragma once - -#include -#include - -namespace armnn -{ - -class RefSoftmaxUint8Workload : public Uint8Workload -{ -public: - using Uint8Workload::Uint8Workload; - virtual void Execute() const override; -}; - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefSoftmaxWorkload.cpp b/src/backends/reference/workloads/RefSoftmaxWorkload.cpp new file mode 100644 index 0000000000..b17666738f --- /dev/null +++ b/src/backends/reference/workloads/RefSoftmaxWorkload.cpp @@ -0,0 +1,39 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#include "RefSoftmaxWorkload.hpp" + +#include "Decoders.hpp" +#include "Encoders.hpp" +#include "RefWorkloadUtils.hpp" +#include "Softmax.hpp" + +#include "Profiling.hpp" + +#include + +namespace armnn +{ + +void RefSoftmaxWorkload::Execute() const +{ + ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefSoftmaxWorkload_Execute"); + + const TensorInfo &inputTensorInfo = GetTensorInfo(m_Data.m_Inputs[0]); + + std::unique_ptr> decoderPtr = MakeDecoder(inputTensorInfo, m_Data.m_Inputs[0]->Map()); + Decoder &decoder = *decoderPtr; + + const TensorInfo &outputTensorInfo = GetTensorInfo(m_Data.m_Outputs[0]); + + std::unique_ptr> encoderPtr = MakeEncoder(outputTensorInfo, m_Data.m_Outputs[0]->Map()); + Encoder &encoder = *encoderPtr; + + Softmax(decoder, + encoder, + inputTensorInfo, + m_Data.m_Parameters.m_Beta); +} +} //namespace armnn diff --git a/src/backends/reference/workloads/RefSoftmaxWorkload.hpp b/src/backends/reference/workloads/RefSoftmaxWorkload.hpp new file mode 100644 index 0000000000..cf3623bf40 --- /dev/null +++ b/src/backends/reference/workloads/RefSoftmaxWorkload.hpp @@ -0,0 +1,21 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#pragma once + +#include +#include + +namespace armnn +{ + +class RefSoftmaxWorkload : public BaseWorkload +{ +public: + using BaseWorkload::BaseWorkload; + virtual void Execute() const override; +}; + +} //namespace armnn diff --git a/src/backends/reference/workloads/RefWorkloads.hpp b/src/backends/reference/workloads/RefWorkloads.hpp index 6ffec2bd06..ab3da88437 100644 --- a/src/backends/reference/workloads/RefWorkloads.hpp +++ b/src/backends/reference/workloads/RefWorkloads.hpp @@ -27,8 +27,7 @@ #include "FullyConnected.hpp" #include "Gather.hpp" #include "RefFloorFloat32Workload.hpp" -#include "RefSoftmaxFloat32Workload.hpp" -#include "RefSoftmaxUint8Workload.hpp" +#include "RefSoftmaxWorkload.hpp" #include "RefResizeBilinearFloat32Workload.hpp" #include "RefBatchNormalizationUint8Workload.hpp" #include "ResizeBilinear.hpp" diff --git a/src/backends/reference/workloads/Softmax.cpp b/src/backends/reference/workloads/Softmax.cpp index 4f1016e86c..6cb219a6cc 100644 --- a/src/backends/reference/workloads/Softmax.cpp +++ b/src/backends/reference/workloads/Softmax.cpp @@ -12,16 +12,19 @@ namespace armnn { /// Computes the softmax function on some inputs, into outputs, with a shape given by tensorInfo. -void Softmax(const float* in, float* out, const TensorInfo& tensorInfo, float beta) +void Softmax(Decoder& in, Encoder& out, const TensorInfo& inputTensorInfo, float beta) { - unsigned int numChannels = tensorInfo.GetShape()[1]; - for (unsigned int n = 0; n < tensorInfo.GetShape()[0]; n++) + unsigned int numChannels = inputTensorInfo.GetShape()[1]; + + for (unsigned int n = 0; n < inputTensorInfo.GetShape()[0]; n++) { // Find maximum channel. - float max = in[n * numChannels]; + in[n * numChannels]; + float max = in.Get(); for (unsigned int c = 1; c < numChannels; c++) { - float val = in[n * numChannels + c]; + in[n * numChannels + c]; + float val = in.Get(); if (val > max) { max = val; @@ -33,7 +36,8 @@ void Softmax(const float* in, float* out, const TensorInfo& tensorInfo, float be float sum = 0.0f; for (unsigned int c = 0; c < numChannels; c++) { - float val = in[n * numChannels + c]; + in[n * numChannels + c]; + float val = in.Get(); exponentials[c] = expf((val - max) * beta); sum += exponentials[c]; } @@ -41,7 +45,8 @@ void Softmax(const float* in, float* out, const TensorInfo& tensorInfo, float be // Divide exponentials by sum to give outputs. for (unsigned int c = 0; c < numChannels; c++) { - out[n * numChannels + c] = exponentials[c] / sum; + out[n * numChannels + c]; + out.Set(exponentials[c] / sum); } } } diff --git a/src/backends/reference/workloads/Softmax.hpp b/src/backends/reference/workloads/Softmax.hpp index 3b974f9e9e..3876293957 100644 --- a/src/backends/reference/workloads/Softmax.hpp +++ b/src/backends/reference/workloads/Softmax.hpp @@ -5,12 +5,13 @@ #pragma once +#include "BaseIterator.hpp" #include namespace armnn { /// Computes the softmax function on some inputs, into outputs, with a shape given by tensorInfo. -void Softmax(const float* in, float* out, const TensorInfo& tensorInfo, float beta); +void Softmax(Decoder& in, Encoder& out, const TensorInfo& inputTensorInfo, float beta); } //namespace armnn -- cgit v1.2.1