diff options
author | Nattapat Chaimanowong <nattapat.chaimanowong@arm.com> | 2019-03-28 09:56:53 +0000 |
---|---|---|
committer | Nattapat Chaimanowong <nattapat.chaimanowong@arm.com> | 2019-03-28 11:18:00 +0000 |
commit | e4294fd7363fd2f5a7be25d1722f20ed935d3445 (patch) | |
tree | b1ebf691cc66e447706081c956a50ce1b4c18e99 /src/armnn | |
parent | 61f54634701d070dbf48e082993fa58104d6f329 (diff) | |
download | armnn-e4294fd7363fd2f5a7be25d1722f20ed935d3445.tar.gz |
IVGCVSW-2874 Add DequantizeLayer and no-op factory method
*Add Dequantize layer to the frontend
*Add Serializer and Deserializer for Dequantize
Change-Id: Ide2647b9e0348d599deb97e61ca4bf66e2f17fc0
Signed-off-by: Nattapat Chaimanowong <nattapat.chaimanowong@arm.com>
Diffstat (limited to 'src/armnn')
-rw-r--r-- | src/armnn/InternalTypes.cpp | 1 | ||||
-rw-r--r-- | src/armnn/InternalTypes.hpp | 1 | ||||
-rw-r--r-- | src/armnn/LayerSupport.cpp | 9 | ||||
-rw-r--r-- | src/armnn/LayersFwd.hpp | 2 | ||||
-rw-r--r-- | src/armnn/Network.cpp | 5 | ||||
-rw-r--r-- | src/armnn/Network.hpp | 2 | ||||
-rw-r--r-- | src/armnn/layers/DequantizeLayer.cpp | 52 | ||||
-rw-r--r-- | src/armnn/layers/DequantizeLayer.hpp | 42 |
8 files changed, 114 insertions, 0 deletions
diff --git a/src/armnn/InternalTypes.cpp b/src/armnn/InternalTypes.cpp index e4a6ac82d5..fe1542b162 100644 --- a/src/armnn/InternalTypes.cpp +++ b/src/armnn/InternalTypes.cpp @@ -24,6 +24,7 @@ char const* GetLayerTypeAsCString(LayerType type) case LayerType::Convolution2d: return "Convolution2d"; case LayerType::Debug: return "Debug"; case LayerType::DepthwiseConvolution2d: return "DepthwiseConvolution2d"; + case LayerType::Dequantize: return "Dequantize"; case LayerType::DetectionPostProcess: return "DetectionPostProcess"; case LayerType::Division: return "Division"; case LayerType::Equal: return "Equal"; diff --git a/src/armnn/InternalTypes.hpp b/src/armnn/InternalTypes.hpp index d6d66031a7..1972e9c1b5 100644 --- a/src/armnn/InternalTypes.hpp +++ b/src/armnn/InternalTypes.hpp @@ -24,6 +24,7 @@ enum class LayerType Convolution2d, Debug, DepthwiseConvolution2d, + Dequantize, DetectionPostProcess, Division, Equal, diff --git a/src/armnn/LayerSupport.cpp b/src/armnn/LayerSupport.cpp index 0e3d2522e4..030973306f 100644 --- a/src/armnn/LayerSupport.cpp +++ b/src/armnn/LayerSupport.cpp @@ -189,6 +189,15 @@ bool IsDepthwiseConvolutionSupported(const BackendId& backend, FORWARD_LAYER_SUPPORT_FUNC(backend, IsDepthwiseConvolutionSupported, input, output, descriptor, weights, biases); } +bool IsDequantizeSupported(const BackendId& backend, + const TensorInfo& input, + const TensorInfo& output, + char* reasonIfUnsupported, + size_t reasonIfUnsupportedMaxLength) +{ + FORWARD_LAYER_SUPPORT_FUNC(backend, IsDequantizeSupported, input, output); +} + bool IsDetectionPostProcessSupported(const BackendId& backend, const TensorInfo& input0, const TensorInfo& input1, diff --git a/src/armnn/LayersFwd.hpp b/src/armnn/LayersFwd.hpp index 9300a75a07..9d87aeeee3 100644 --- a/src/armnn/LayersFwd.hpp +++ b/src/armnn/LayersFwd.hpp @@ -16,6 +16,7 @@ #include "layers/Convolution2dLayer.hpp" #include "layers/DebugLayer.hpp" #include "layers/DepthwiseConvolution2dLayer.hpp" +#include "layers/DequantizeLayer.hpp" #include "layers/DetectionPostProcessLayer.hpp" #include "layers/DivisionLayer.hpp" #include "layers/EqualLayer.hpp" @@ -86,6 +87,7 @@ DECLARE_LAYER(ConvertFp32ToFp16) DECLARE_LAYER(Convolution2d) DECLARE_LAYER(Debug) DECLARE_LAYER(DepthwiseConvolution2d) +DECLARE_LAYER(Dequantize) DECLARE_LAYER(DetectionPostProcess) DECLARE_LAYER(Division) DECLARE_LAYER(Equal) diff --git a/src/armnn/Network.cpp b/src/armnn/Network.cpp index c5dfbd75ec..6dbd4611df 100644 --- a/src/armnn/Network.cpp +++ b/src/armnn/Network.cpp @@ -935,6 +935,11 @@ IConnectableLayer *Network::AddQuantizeLayer(const char *name) return m_Graph->AddLayer<QuantizeLayer>(name); } +IConnectableLayer* Network::AddDequantizeLayer(const char* name) +{ + return m_Graph->AddLayer<DequantizeLayer>(name); +} + IConnectableLayer* Network::AddStridedSliceLayer(const StridedSliceDescriptor& stridedSliceDescriptor, const char* name) { diff --git a/src/armnn/Network.hpp b/src/armnn/Network.hpp index 5ed8cca2f2..782531acde 100644 --- a/src/armnn/Network.hpp +++ b/src/armnn/Network.hpp @@ -72,6 +72,8 @@ public: const ConstTensor& biases, const char* name = nullptr) override; + IConnectableLayer* AddDequantizeLayer(const char* name = nullptr) override; + IConnectableLayer* AddDetectionPostProcessLayer( const DetectionPostProcessDescriptor& descriptor, const ConstTensor& anchors, diff --git a/src/armnn/layers/DequantizeLayer.cpp b/src/armnn/layers/DequantizeLayer.cpp new file mode 100644 index 0000000000..4dd30de77b --- /dev/null +++ b/src/armnn/layers/DequantizeLayer.cpp @@ -0,0 +1,52 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// +#include "DequantizeLayer.hpp" + +#include "LayerCloneBase.hpp" + +#include <backendsCommon/WorkloadData.hpp> +#include <backendsCommon/WorkloadFactory.hpp> + +namespace armnn +{ + +DequantizeLayer::DequantizeLayer(const char* name) + : Layer(1, 1, LayerType::Dequantize, name) +{} + +std::unique_ptr<IWorkload> DequantizeLayer::CreateWorkload(const Graph& graph, + const IWorkloadFactory& factory) const +{ + DequantizeQueueDescriptor descriptor; + + return factory.CreateDequantize(descriptor, PrepInfoAndDesc(descriptor, graph)); +} + +DequantizeLayer* DequantizeLayer::Clone(Graph& graph) const +{ + return CloneBase<DequantizeLayer>(graph, GetName()); +} + +void DequantizeLayer::ValidateTensorShapesFromInputs() +{ + VerifyLayerConnections(1, CHECK_LOCATION()); + + std::vector<TensorShape> inferredShapes = InferOutputShapes({ + GetInputSlot(0).GetConnection()->GetTensorInfo().GetShape() }); + + BOOST_ASSERT(inferredShapes.size() == 1); + + ConditionalThrowIfNotEqual<LayerValidationException>( + "DequantizeLayer: TensorShape set on OutputSlot[0] does not match the inferred shape.", + GetOutputSlot(0).GetTensorInfo().GetShape(), + inferredShapes[0]); +} + +void DequantizeLayer::Accept(ILayerVisitor& visitor) const +{ + visitor.VisitDequantizeLayer(this, GetName()); +} + +} // namespace armnn diff --git a/src/armnn/layers/DequantizeLayer.hpp b/src/armnn/layers/DequantizeLayer.hpp new file mode 100644 index 0000000000..1340f96a27 --- /dev/null +++ b/src/armnn/layers/DequantizeLayer.hpp @@ -0,0 +1,42 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// +#pragma once + +#include "Layer.hpp" + +namespace armnn +{ + +/// This layer dequantizes the input tensor. +class DequantizeLayer : public Layer +{ +public: + /// Makes a workload for the Dequantize type. + /// @param [in] graph The graph where this layer can be found. + /// @param [in] factory The workload factory which will create the workload. + /// @return A pointer to the created workload, or nullptr if not created. + virtual std::unique_ptr<IWorkload> CreateWorkload(const Graph& graph, + const IWorkloadFactory& factory) const override; + + /// Creates a dynamically-allocated copy of this layer. + /// @param [in] graph The graph into which this layer is being cloned. + DequantizeLayer* Clone(Graph& graph) const override; + + /// Check if the input tensor shape(s) + /// will lead to a valid configuration of @ref DequantizeLayer. + void ValidateTensorShapesFromInputs() override; + + void Accept(ILayerVisitor& visitor) const override; + +protected: + /// Constructor to create a DequantizeLayer. + /// @param [in] name Optional name for the layer. + DequantizeLayer(const char* name); + + /// Default destructor + ~DequantizeLayer() = default; +}; + +} // namespace armnn |