diff options
Diffstat (limited to 'src/backends/neon/workloads')
-rw-r--r-- | src/backends/neon/workloads/CMakeLists.txt | 6 | ||||
-rw-r--r-- | src/backends/neon/workloads/NeonReshapeFloatWorkload.cpp | 32 | ||||
-rw-r--r-- | src/backends/neon/workloads/NeonReshapeFloatWorkload.hpp | 29 | ||||
-rw-r--r-- | src/backends/neon/workloads/NeonReshapeUint8Workload.cpp | 30 | ||||
-rw-r--r-- | src/backends/neon/workloads/NeonReshapeWorkload.cpp | 29 | ||||
-rw-r--r-- | src/backends/neon/workloads/NeonReshapeWorkload.hpp (renamed from src/backends/neon/workloads/NeonReshapeUint8Workload.hpp) | 9 | ||||
-rw-r--r-- | src/backends/neon/workloads/NeonWorkloads.hpp | 3 |
7 files changed, 35 insertions, 103 deletions
diff --git a/src/backends/neon/workloads/CMakeLists.txt b/src/backends/neon/workloads/CMakeLists.txt index 06eb504e99..7c2247256d 100644 --- a/src/backends/neon/workloads/CMakeLists.txt +++ b/src/backends/neon/workloads/CMakeLists.txt @@ -53,10 +53,8 @@ list(APPEND armnnNeonBackendWorkloads_sources NeonPooling2dFloatWorkload.hpp NeonPooling2dUint8Workload.cpp NeonPooling2dUint8Workload.hpp - NeonReshapeFloatWorkload.cpp - NeonReshapeFloatWorkload.hpp - NeonReshapeUint8Workload.cpp - NeonReshapeUint8Workload.hpp + NeonReshapeWorkload.cpp + NeonReshapeWorkload.hpp NeonSoftmaxBaseWorkload.cpp NeonSoftmaxBaseWorkload.hpp NeonSoftmaxFloatWorkload.cpp diff --git a/src/backends/neon/workloads/NeonReshapeFloatWorkload.cpp b/src/backends/neon/workloads/NeonReshapeFloatWorkload.cpp deleted file mode 100644 index 2dae9466bb..0000000000 --- a/src/backends/neon/workloads/NeonReshapeFloatWorkload.cpp +++ /dev/null @@ -1,32 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#include "NeonReshapeFloatWorkload.hpp" - - - -namespace armnn -{ - -NeonReshapeFloatWorkload::NeonReshapeFloatWorkload(const ReshapeQueueDescriptor& descriptor, - const WorkloadInfo& info) - : FloatWorkload<ReshapeQueueDescriptor>(descriptor, info) -{ - m_Data.ValidateInputsOutputs("NeonReshapeFloatWorkload", 1, 1); - - arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor(); - arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor(); - - m_Layer.configure(&input, &output); -} - -void NeonReshapeFloatWorkload::Execute() const -{ - ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonReshapeFloatWorkload_Execute"); - m_Layer.run(); -} - -} //namespace armnn - diff --git a/src/backends/neon/workloads/NeonReshapeFloatWorkload.hpp b/src/backends/neon/workloads/NeonReshapeFloatWorkload.hpp deleted file mode 100644 index bdef862419..0000000000 --- a/src/backends/neon/workloads/NeonReshapeFloatWorkload.hpp +++ /dev/null @@ -1,29 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#pragma once - -#include <backends/neon/workloads/NeonWorkloadUtils.hpp> - -namespace armnn -{ - -class NeonReshapeFloatWorkload : public FloatWorkload<ReshapeQueueDescriptor> -{ -public: - NeonReshapeFloatWorkload(const ReshapeQueueDescriptor& descriptor, const WorkloadInfo& info); - - virtual void Execute() const override; - -private: - mutable arm_compute::NEReshapeLayer m_Layer; -}; - -} //namespace armnn - - - - - diff --git a/src/backends/neon/workloads/NeonReshapeUint8Workload.cpp b/src/backends/neon/workloads/NeonReshapeUint8Workload.cpp deleted file mode 100644 index 41aa07fe49..0000000000 --- a/src/backends/neon/workloads/NeonReshapeUint8Workload.cpp +++ /dev/null @@ -1,30 +0,0 @@ -// -// Copyright © 2017 Arm Ltd. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#include "NeonReshapeUint8Workload.hpp" - - - - -namespace armnn -{ -NeonReshapeUint8Workload::NeonReshapeUint8Workload(const ReshapeQueueDescriptor& descriptor, - const WorkloadInfo& info) - : Uint8Workload<ReshapeQueueDescriptor>(descriptor, info) -{ - m_Data.ValidateInputsOutputs("NeonReshapeUint8Workload", 1, 1); - - arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor(); - arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor(); - - m_Layer.configure(&input, &output); -} - -void NeonReshapeUint8Workload::Execute() const -{ - ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonReshapeUint8Workload_Execute"); - m_Layer.run(); -} -} //namespace armnn diff --git a/src/backends/neon/workloads/NeonReshapeWorkload.cpp b/src/backends/neon/workloads/NeonReshapeWorkload.cpp new file mode 100644 index 0000000000..c2dcdd500c --- /dev/null +++ b/src/backends/neon/workloads/NeonReshapeWorkload.cpp @@ -0,0 +1,29 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#include "NeonReshapeWorkload.hpp" + +namespace armnn +{ + +NeonReshapeWorkload::NeonReshapeWorkload(const ReshapeQueueDescriptor& descriptor, + const WorkloadInfo& info) + : BaseWorkload<ReshapeQueueDescriptor>(descriptor, info) +{ + m_Data.ValidateInputsOutputs("NeonReshapeWorkload", 1, 1); + + arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor(); + arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor(); + + m_Layer.configure(&input, &output); +} + +void NeonReshapeWorkload::Execute() const +{ + ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonReshapeWorkload_Execute"); + m_Layer.run(); +} + +} //namespace armnn diff --git a/src/backends/neon/workloads/NeonReshapeUint8Workload.hpp b/src/backends/neon/workloads/NeonReshapeWorkload.hpp index 4951873f0b..8fd278b2eb 100644 --- a/src/backends/neon/workloads/NeonReshapeUint8Workload.hpp +++ b/src/backends/neon/workloads/NeonReshapeWorkload.hpp @@ -10,10 +10,11 @@ namespace armnn { -class NeonReshapeUint8Workload : public Uint8Workload<ReshapeQueueDescriptor> +class NeonReshapeWorkload : public BaseWorkload<ReshapeQueueDescriptor> { public: - NeonReshapeUint8Workload(const ReshapeQueueDescriptor& descriptor, const WorkloadInfo& info); + NeonReshapeWorkload(const ReshapeQueueDescriptor& descriptor, const WorkloadInfo& info); + virtual void Execute() const override; private: @@ -21,7 +22,3 @@ private: }; } //namespace armnn - - - - diff --git a/src/backends/neon/workloads/NeonWorkloads.hpp b/src/backends/neon/workloads/NeonWorkloads.hpp index d5a76b44cc..c877c943f4 100644 --- a/src/backends/neon/workloads/NeonWorkloads.hpp +++ b/src/backends/neon/workloads/NeonWorkloads.hpp @@ -29,8 +29,7 @@ #include "NeonPooling2dBaseWorkload.hpp" #include "NeonPooling2dFloatWorkload.hpp" #include "NeonPooling2dUint8Workload.hpp" -#include "NeonReshapeFloatWorkload.hpp" -#include "NeonReshapeUint8Workload.hpp" +#include "NeonReshapeWorkload.hpp" #include "NeonSoftmaxFloatWorkload.hpp" #include "NeonSoftmaxUint8Workload.hpp" #include "NeonSplitterFloatWorkload.hpp" |