diff options
author | Cathal Corbett <cathal.corbett@arm.com> | 2022-05-16 15:20:56 +0100 |
---|---|---|
committer | Ryan OShea <ryan.oshea3@arm.com> | 2022-05-19 11:07:09 +0100 |
commit | 630ce65543c08d8e7fca5be80f9a64122744d135 (patch) | |
tree | f78db6d55bc9feaf884d83212c9ec9195f482225 /src/armnn/optimizations/ConvertConstPermuteLayersToConstLayers.hpp | |
parent | 5841c740ba6bc6c8c3e96d24156dc47907af6430 (diff) | |
download | armnn-630ce65543c08d8e7fca5be80f9a64122744d135.tar.gz |
IVGCVSW-6147 ConstTensorsAsInput: Optimizer - FusePermuteIntoConstLayer
* No trailing permute layer after a constant layer
* Unit test for optimization
Signed-off-by: Cathal Corbett <cathal.corbett@arm.com>
Change-Id: I0d098f5af41d2c55df7cef1ccfb848093320ddc1
Diffstat (limited to 'src/armnn/optimizations/ConvertConstPermuteLayersToConstLayers.hpp')
-rw-r--r-- | src/armnn/optimizations/ConvertConstPermuteLayersToConstLayers.hpp | 127 |
1 files changed, 127 insertions, 0 deletions
diff --git a/src/armnn/optimizations/ConvertConstPermuteLayersToConstLayers.hpp b/src/armnn/optimizations/ConvertConstPermuteLayersToConstLayers.hpp new file mode 100644 index 0000000000..2cc3e8eaef --- /dev/null +++ b/src/armnn/optimizations/ConvertConstPermuteLayersToConstLayers.hpp @@ -0,0 +1,127 @@ +// +// Copyright © 2022 Arm Ltd and Contributors. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#pragma once + +#include "Optimization.hpp" +#include <armnnUtils/Permute.hpp> +#include <ResolveType.hpp> + +namespace armnn +{ +namespace optimizations +{ + +class ConvertConstPermuteLayersToConstLayers +{ +public: + void Run(Graph& graph, InputSlot& connection) const + { + Layer& base = connection.GetConnectedOutputSlot()->GetOwningLayer(); + Layer& child = connection.GetOwningLayer(); + + ARMNN_ASSERT(base.GetType() == LayerType::Constant); + ARMNN_ASSERT(child.GetType() == LayerType::Permute); + + if (base.GetDataType() == child.GetDataType()) + { + switch (base.GetDataType()) + { + case DataType::Float16: + ReplaceConstPermuteLayer<DataType::Float16>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + case DataType::Float32: + ReplaceConstPermuteLayer<DataType::Float32>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + case DataType::QAsymmU8: + ReplaceConstPermuteLayer<DataType::QAsymmU8>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + case DataType::Signed32: + ReplaceConstPermuteLayer<DataType::Signed32>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + case DataType::QSymmS16: + ReplaceConstPermuteLayer<DataType::QSymmS16>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + case DataType::QSymmS8: + ReplaceConstPermuteLayer<DataType::QSymmS8>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + case DataType::QAsymmS8: + ReplaceConstPermuteLayer<DataType::QAsymmS8>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + case DataType::BFloat16: + ReplaceConstPermuteLayer<DataType::BFloat16>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + case DataType::Signed64: + ReplaceConstPermuteLayer<DataType::Signed64>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + case DataType::Boolean: + ReplaceConstPermuteLayer<DataType::Boolean>(graph, + PolymorphicDowncast<ConstantLayer*>(&base), + PolymorphicDowncast<PermuteLayer*>(&child)); + break; + } + } + } +protected: + ConvertConstPermuteLayersToConstLayers() = default; + ~ConvertConstPermuteLayersToConstLayers() = default; +private: + template<armnn::DataType ArmnnType, + typename T = armnn::ResolveType<ArmnnType>> + static void ReplaceConstPermuteLayer(Graph& graph, + ConstantLayer* constantLayer, + PermuteLayer* permuteLayer) + { + IgnoreUnused(graph); + /** + * This optimisation is to find situations where a constant set of inputs is being provided to a Permute + * layer. In this case we don't want the overhead of Permuting the values on every inference, instead we + * want to Permute them once and store them in a Const layer to be used everytime as they will not change. + */ + TensorInfo outputPermuteInfo = permuteLayer->GetOutputSlot(0).GetTensorInfo(); + std::vector<T> newValues(outputPermuteInfo.GetNumElements()); + armnnUtils::Permute(outputPermuteInfo.GetShape(), permuteLayer->GetPermutation(), + constantLayer->m_LayerOutput->Map(true), newValues.data(), + GetDataTypeSize(outputPermuteInfo.GetDataType())); + + TensorInfo newInfo = outputPermuteInfo; + newInfo.SetConstant(true); + ConstTensor newInput(newInfo, newValues); + constantLayer->m_LayerOutput.reset(new ScopedTensorHandle(newInput)); + + // Moves connections in permute output to the constant layer. + // Permute layer will be removed if left unconnected. + permuteLayer->GetOutputSlot().MoveAllConnections(constantLayer->GetOutputSlot()); + + // Updating the output tensor + constantLayer->GetOutputSlot(0).SetTensorInfo(newInfo); + ARMNN_ASSERT(constantLayer->GetOutputSlot(0).GetTensorInfo().IsConstant() == true); + } +}; + +using FusePermuteIntoConstLayer = OptimizeForConnection<ConstantLayer, + PermuteLayer, + ConvertConstPermuteLayersToConstLayers>; + +} // namespace optimizations +} // namespace armnn
\ No newline at end of file |