diff options
author | Kevin May <kevin.may@arm.com> | 2021-04-22 16:51:18 +0100 |
---|---|---|
committer | Kevin May <kevin.may@arm.com> | 2021-04-27 15:02:45 +0000 |
commit | 4a621c43174b6bdd9dc0bff839b245bc2139d6a6 (patch) | |
tree | 88b2bbc803284e7c5051c8812a468e5f3c834b9c /src/armnn/QuantizerStrategy.hpp | |
parent | 65542e4f250a66fd0aba0c1b509e11cb518ee2d5 (diff) | |
download | armnn-4a621c43174b6bdd9dc0bff839b245bc2139d6a6.tar.gz |
IVGCVSW-5719 Remove Quantizer
Signed-off-by: Kevin May <kevin.may@arm.com>
Change-Id: I8a29df03afdd6f1cc8413b219e179272cd2d51cf
Diffstat (limited to 'src/armnn/QuantizerStrategy.hpp')
-rw-r--r-- | src/armnn/QuantizerStrategy.hpp | 63 |
1 files changed, 0 insertions, 63 deletions
diff --git a/src/armnn/QuantizerStrategy.hpp b/src/armnn/QuantizerStrategy.hpp deleted file mode 100644 index f782959020..0000000000 --- a/src/armnn/QuantizerStrategy.hpp +++ /dev/null @@ -1,63 +0,0 @@ -// -// Copyright © 2021 Arm Ltd and Contributors. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#include "Network.hpp" -#include "NetworkQuantizerUtils.hpp" -#include "StaticRangeStrategy.hpp" - -#include <armnn/utility/NumericCast.hpp> -#include <armnn/utility/PolymorphicDowncast.hpp> - -namespace armnn -{ -class QuantizerStrategy : public IStrategy -{ -public : - QuantizerStrategy(const RangeTracker& rangeTracker, - const IQuantizationScheme* quantizationScheme, - bool preserveType); - - ~QuantizerStrategy() = default; - - void ExecuteStrategy(const armnn::IConnectableLayer* layer, - const BaseDescriptor& descriptor, - const std::vector<armnn::ConstTensor>& constants, - const char* name, - const armnn::LayerBindingId id) override; - - /// Extract the quantized network - INetworkPtr RetrieveFinalNetwork() { return std::move(m_QuantizedNetwork); } - -private: - /// Connects the layer to preceeding layers and sets the quantization parameters based on recorded ranges - void SetQuantizedInputConnections(const IConnectableLayer* srcLayer, IConnectableLayer* quantizedLayer); - - /// Record the guids so we can easily find the layers later - void RecordLayer(const IConnectableLayer* srcLayer, IConnectableLayer* qLayer); - - /// Sets the bias quantization scale based on input and weight scales - ConstTensor CreateQuantizedBias(const IConnectableLayer* srcLayer, - const ConstTensor& weights, - const Optional<ConstTensor>& biases, - std::vector<int32_t>& weightsBacking); - - /// Reference to the static range visitor used to retrieve the quantization ranges - const RangeTracker& m_Ranges; - - /// Quantized version of the model we are building up - INetworkPtr m_QuantizedNetwork; - - /// Mapping from input network guids to quantized network guids - std::unordered_map<LayerGuid, LayerGuid> m_OriginalToQuantizedGuidMap; - - /// Mapping from guid to layer in quantized network - std::unordered_map<LayerGuid, IConnectableLayer*> m_QuantizedGuidToLayerMap; - - const IQuantizationScheme* m_QuantizationScheme; - - const bool m_PreserveType; -}; - -} //namespace armnn
\ No newline at end of file |