aboutsummaryrefslogtreecommitdiff
path: root/src/backends/NeonWorkloads
diff options
context:
space:
mode:
authorkevmay01 <kevin.may@arm.com>2018-09-26 10:21:55 +0100
committerMatthew Bentham <matthew.bentham@arm.com>2018-10-10 16:16:57 +0100
commite448be3ac55897a3eabe85962891f8414f8e3cf9 (patch)
tree11e46d4979eb6d8d70c7f43d5cf690fc6f78d605 /src/backends/NeonWorkloads
parent9fc824a596d6bddd27e5aa1438b115e71a117aa3 (diff)
downloadarmnn-e448be3ac55897a3eabe85962891f8414f8e3cf9.tar.gz
IVGCVSW-1927 Add Neon 8-bit FullyConnected support
Change-Id: Idf4cc7a9a7d3261b9eceb653b999257506cdae76
Diffstat (limited to 'src/backends/NeonWorkloads')
-rw-r--r--src/backends/NeonWorkloads/CMakeLists.txt4
-rw-r--r--src/backends/NeonWorkloads/NeonFullyConnectedWorkload.cpp (renamed from src/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.cpp)32
-rw-r--r--src/backends/NeonWorkloads/NeonFullyConnectedWorkload.hpp (renamed from src/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.hpp)6
-rw-r--r--src/backends/NeonWorkloads/backend.mk2
4 files changed, 29 insertions, 15 deletions
diff --git a/src/backends/NeonWorkloads/CMakeLists.txt b/src/backends/NeonWorkloads/CMakeLists.txt
index 02cb53dff1..ca44bcc1e1 100644
--- a/src/backends/NeonWorkloads/CMakeLists.txt
+++ b/src/backends/NeonWorkloads/CMakeLists.txt
@@ -37,8 +37,8 @@ list(APPEND armnnNeonBackend_sources
NeonDepthwiseConvolutionUint8Workload.hpp
NeonFloorFloatWorkload.cpp
NeonFloorFloatWorkload.hpp
- NeonFullyConnectedFloatWorkload.cpp
- NeonFullyConnectedFloatWorkload.hpp
+ NeonFullyConnectedWorkload.cpp
+ NeonFullyConnectedWorkload.hpp
NeonL2NormalizationFloatWorkload.cpp
NeonL2NormalizationFloatWorkload.hpp
NeonLstmFloatWorkload.cpp
diff --git a/src/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.cpp b/src/backends/NeonWorkloads/NeonFullyConnectedWorkload.cpp
index 8fcc0956e2..8cebb4f48f 100644
--- a/src/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.cpp
+++ b/src/backends/NeonWorkloads/NeonFullyConnectedWorkload.cpp
@@ -3,7 +3,7 @@
// SPDX-License-Identifier: MIT
//
-#include "NeonFullyConnectedFloatWorkload.hpp"
+#include "NeonFullyConnectedWorkload.hpp"
#include <backends/aclCommon/ArmComputeTensorUtils.hpp>
#include <backends/aclCommon/ArmComputeUtils.hpp>
@@ -42,12 +42,12 @@ arm_compute::Status NeonFullyConnectedWorkloadValidate(const TensorInfo& input,
fullyConnectedLayerInfo);
}
-NeonFullyConnectedFloatWorkload::NeonFullyConnectedFloatWorkload(const FullyConnectedQueueDescriptor& descriptor,
+NeonFullyConnectedWorkload::NeonFullyConnectedWorkload(const FullyConnectedQueueDescriptor& descriptor,
const WorkloadInfo& info, std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
- : FloatWorkload<FullyConnectedQueueDescriptor>(descriptor, info)
+ : BaseWorkload<FullyConnectedQueueDescriptor>(descriptor, info)
, m_FullyConnectedLayer(memoryManager)
{
- m_Data.ValidateInputsOutputs("NeonFullyConnectedFloatWorkload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonFullyConnectedWorkload", 1, 1);
arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -67,11 +67,25 @@ NeonFullyConnectedFloatWorkload::NeonFullyConnectedFloatWorkload(const FullyConn
m_FullyConnectedLayer.configure(&input, m_WeightsTensor.get(), m_BiasesTensor.get(), &output, fc_info);
// Allocate
- InitializeArmComputeTensorDataForFloatTypes(*m_WeightsTensor, m_Data.m_Weight);
+ if (m_Data.m_Weight->GetTensorInfo().GetDataType() == DataType::QuantisedAsymm8)
+ {
+ InitialiseArmComputeTensorData(*m_WeightsTensor, m_Data.m_Weight->GetConstTensor<uint8_t>());
+ }
+ else
+ {
+ InitializeArmComputeTensorDataForFloatTypes(*m_WeightsTensor, m_Data.m_Weight);
+ }
if (m_BiasesTensor)
{
- InitializeArmComputeTensorDataForFloatTypes(*m_BiasesTensor, m_Data.m_Bias);
+ if (m_Data.m_Bias->GetTensorInfo().GetDataType() == DataType::Signed32)
+ {
+ InitialiseArmComputeTensorData(*m_BiasesTensor, m_Data.m_Bias->GetConstTensor<int32_t>());
+ }
+ else
+ {
+ InitializeArmComputeTensorDataForFloatTypes(*m_BiasesTensor, m_Data.m_Bias);
+ }
}
// Force Compute Library to perform the necessary copying and reshaping, after which
@@ -80,13 +94,13 @@ NeonFullyConnectedFloatWorkload::NeonFullyConnectedFloatWorkload(const FullyConn
FreeUnusedTensors();
}
-void NeonFullyConnectedFloatWorkload::Execute() const
+void NeonFullyConnectedWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonFullyConnectedFloatWorkload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonFullyConnectedWorkload_Execute");
m_FullyConnectedLayer.run();
}
-void NeonFullyConnectedFloatWorkload::FreeUnusedTensors()
+void NeonFullyConnectedWorkload::FreeUnusedTensors()
{
FreeTensorIfUnused(m_WeightsTensor);
FreeTensorIfUnused(m_BiasesTensor);
diff --git a/src/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.hpp b/src/backends/NeonWorkloads/NeonFullyConnectedWorkload.hpp
index 27e5717b04..11991f87b5 100644
--- a/src/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.hpp
+++ b/src/backends/NeonWorkloads/NeonFullyConnectedWorkload.hpp
@@ -20,11 +20,11 @@ arm_compute::Status NeonFullyConnectedWorkloadValidate(const TensorInfo& input,
const TensorInfo& biases,
const FullyConnectedDescriptor& descriptor);
-class NeonFullyConnectedFloatWorkload : public FloatWorkload<FullyConnectedQueueDescriptor>
+class NeonFullyConnectedWorkload : public BaseWorkload<FullyConnectedQueueDescriptor>
{
public:
- NeonFullyConnectedFloatWorkload(const FullyConnectedQueueDescriptor& descriptor, const WorkloadInfo& info,
- std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
+ NeonFullyConnectedWorkload(const FullyConnectedQueueDescriptor& descriptor, const WorkloadInfo& info,
+ std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
virtual void Execute() const override;
private:
diff --git a/src/backends/NeonWorkloads/backend.mk b/src/backends/NeonWorkloads/backend.mk
index 06525c9df2..4c7fbe6fe0 100644
--- a/src/backends/NeonWorkloads/backend.mk
+++ b/src/backends/NeonWorkloads/backend.mk
@@ -19,7 +19,7 @@ BACKEND_SOURCES := \
NeonDepthwiseConvolutionFloatWorkload.cpp \
NeonDepthwiseConvolutionUint8Workload.cpp \
NeonFloorFloatWorkload.cpp \
- NeonFullyConnectedFloatWorkload.cpp \
+ NeonFullyConnectedWorkload.cpp \
NeonL2NormalizationFloatWorkload.cpp \
NeonLstmFloatWorkload.cpp \
NeonMergerFloatWorkload.cpp \