aboutsummaryrefslogtreecommitdiff
path: root/src/armnn/DynamicQuantizationVisitor.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/armnn/DynamicQuantizationVisitor.cpp')
-rw-r--r--src/armnn/DynamicQuantizationVisitor.cpp330
1 files changed, 330 insertions, 0 deletions
diff --git a/src/armnn/DynamicQuantizationVisitor.cpp b/src/armnn/DynamicQuantizationVisitor.cpp
new file mode 100644
index 0000000000..9b33fb7642
--- /dev/null
+++ b/src/armnn/DynamicQuantizationVisitor.cpp
@@ -0,0 +1,330 @@
+//
+// Copyright © 2017 Arm Ltd. All rights reserved.
+// SPDX-License-Identifier: MIT
+//
+
+#include "DynamicQuantizationVisitor.hpp"
+#include "NetworkUtils.hpp"
+
+#include <boost/core/ignore_unused.hpp>
+#include <armnn/Descriptors.hpp>
+#include <armnn/Types.hpp>
+
+#include <limits>
+
+namespace armnn
+{
+
+DynamicQuantizationVisitor::DynamicQuantizationVisitor(RangeTracker& rangeTracker, Graph& graph)
+ : m_RangeTracker(rangeTracker),
+ m_Graph(graph)
+{}
+
+void DynamicQuantizationVisitor::SetRange(const IConnectableLayer* layer, unsigned int outputIdx, float min, float max)
+{
+ m_RangeTracker.SetRange(layer, outputIdx, min, max);
+}
+
+void DynamicQuantizationVisitor::ForwardParentParameters(const IConnectableLayer* layer)
+{
+ for (unsigned int i = 0; i < layer->GetNumInputSlots(); ++i)
+ {
+ const IOutputSlot *outputSlot = layer->GetInputSlot(i).GetConnection();
+ LayerGuid previousLayerId = outputSlot->GetOwningLayerGuid();
+ unsigned int ownerIndex = outputSlot->CalculateIndexOnOwner();
+ const auto parentRange = m_RangeTracker.GetRange(previousLayerId, ownerIndex);
+ SetRange(layer, i, parentRange.first, parentRange.second);
+ }
+}
+
+void DynamicQuantizationVisitor::AddToCalibratedLayers(const IConnectableLayer* layer)
+{
+ m_LayersToCalibrate.push_back(layer);
+}
+
+void DynamicQuantizationVisitor::AddToNonCalibratedLayers(const IConnectableLayer* layer)
+{
+ m_LayersNotToCalibrate.push_back(layer);
+}
+
+void DynamicQuantizationVisitor::FinishVisit()
+{
+ for (const IConnectableLayer* layer : m_LayersToCalibrate)
+ {
+ std::vector<DebugLayer*> newDebugLayers = InsertDebugLayerAfter(
+ m_Graph, *boost::polymorphic_downcast<Layer*>(const_cast<IConnectableLayer*>(layer)));
+ // record them so we can take them out again efficiently afterward
+ m_DebugLayers.insert(std::end(m_DebugLayers), std::begin(newDebugLayers), std::end(newDebugLayers));
+ }
+}
+
+void DynamicQuantizationVisitor::RemoveDebugLayers()
+{
+ for (DebugLayer* debugLayer : m_DebugLayers)
+ {
+ OutputSlot& proceedingOutputSlot = *debugLayer->GetInputSlot(0).GetConnectedOutputSlot();
+ InputSlot& succeedingInputSlot = *debugLayer->GetOutputSlot(0).GetConnection(0);
+ proceedingOutputSlot.Disconnect(debugLayer->GetInputSlot(0));
+ debugLayer->GetOutputSlot(0).Disconnect(succeedingInputSlot);
+
+ m_Graph.EraseLayer(debugLayer);
+
+ proceedingOutputSlot.Connect(succeedingInputSlot);
+ }
+ m_DebugLayers.clear();
+}
+
+void DynamicQuantizationVisitor::VisitNonCalibratedLayers() {
+ RemoveDebugLayers();
+ for (const IConnectableLayer* layer : m_LayersNotToCalibrate)
+ {
+ ForwardParentParameters(layer);
+ }
+}
+
+void DynamicQuantizationVisitor::VisitAdditionLayer(const IConnectableLayer* layer, const char* name)
+{
+ SetRange(layer, 0, -20.f, 20.f);
+ AddToCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitBatchNormalizationLayer(const IConnectableLayer* layer,
+ const BatchNormalizationDescriptor& desc,
+ const ConstTensor& mean,
+ const ConstTensor& variance,
+ const ConstTensor& beta,
+ const ConstTensor& gamma,
+ const char* name)
+{
+ boost::ignore_unused(desc);
+ boost::ignore_unused(mean);
+ boost::ignore_unused(variance);
+ boost::ignore_unused(beta);
+ boost::ignore_unused(gamma);
+ boost::ignore_unused(name);
+ SetRange(layer, 0, -15.0f, 15.0f);
+ AddToCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitConvolution2dLayer(const IConnectableLayer* layer,
+ const Convolution2dDescriptor& convolution2dDescriptor,
+ const ConstTensor& weights,
+ const Optional<ConstTensor>& biases,
+ const char* name)
+{
+ boost::ignore_unused(convolution2dDescriptor);
+ boost::ignore_unused(weights);
+ boost::ignore_unused(biases);
+ boost::ignore_unused(name);
+ SetRange(layer, 0, -15.0f, 15.0f);
+ AddToCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitDepthwiseConvolution2dLayer(const IConnectableLayer* layer,
+ const DepthwiseConvolution2dDescriptor& desc,
+ const ConstTensor& weights,
+ const Optional<ConstTensor>& biases,
+ const char* name)
+{
+ boost::ignore_unused(desc);
+ boost::ignore_unused(weights);
+ boost::ignore_unused(biases);
+ boost::ignore_unused(name);
+ SetRange(layer, 0, -15.0f, 15.0f);
+ AddToCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitActivationLayer(const IConnectableLayer* layer,
+ const ActivationDescriptor& activationDescriptor,
+ const char* name)
+{
+ switch (activationDescriptor.m_Function)
+ {
+ // Range is 0, 15 for Abs, Linear, ReLu and Soft ReLu
+ case ActivationFunction::Abs:
+ case ActivationFunction::Linear:
+ case ActivationFunction::ReLu:
+ case ActivationFunction::SoftReLu:
+ SetRange(layer, 0, 0.f, 15.f);
+ break;
+ case ActivationFunction::BoundedReLu:
+ SetRange(layer, 0, 0.f, activationDescriptor.m_A);
+ break;
+ case ActivationFunction::TanH:
+ SetRange(layer, 0, -1.f, 1.f);
+ break;
+ case ActivationFunction::LeakyReLu:
+ SetRange(layer, 0, -5.f, 15.f);
+ break;
+ default:
+ SetRange(layer, 0, -15.f, 15.f);
+ break;
+ }
+ AddToCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitFullyConnectedLayer(const IConnectableLayer *layer,
+ const FullyConnectedDescriptor& desc,
+ const ConstTensor& weights,
+ const Optional<ConstTensor>& biases,
+ const char *name)
+{
+ boost::ignore_unused(desc);
+ boost::ignore_unused(weights);
+ boost::ignore_unused(biases);
+ boost::ignore_unused(name);
+ SetRange(layer, 0, -15.0f, 15.0f);
+ AddToCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitPermuteLayer(const IConnectableLayer* layer,
+ const PermuteDescriptor& permuteDescriptor,
+ const char* name)
+{
+ boost::ignore_unused(permuteDescriptor);
+ boost::ignore_unused(name);
+ AddToNonCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitSpaceToBatchNdLayer(const IConnectableLayer* layer,
+ const SpaceToBatchNdDescriptor& spaceToBatchNdDescriptor,
+ const char* name)
+{
+ boost::ignore_unused(spaceToBatchNdDescriptor);
+ boost::ignore_unused(name);
+ AddToNonCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitPooling2dLayer(const IConnectableLayer* layer,
+ const Pooling2dDescriptor& pooling2dDescriptor,
+ const char* name)
+{
+ boost::ignore_unused(pooling2dDescriptor);
+ boost::ignore_unused(name);
+ AddToNonCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitSoftmaxLayer(const IConnectableLayer* layer,
+ const SoftmaxDescriptor& softmaxDescriptor,
+ const char* name)
+{
+ boost::ignore_unused(softmaxDescriptor);
+ boost::ignore_unused(name);
+ SetRange(layer, 0, 0.f, 1.f);
+ AddToCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitConstantLayer(const IConnectableLayer* layer,
+ const ConstTensor& input,
+ const char* name)
+{
+ boost::ignore_unused(name);
+
+ if (input.GetDataType() != DataType::Float32)
+ {
+ throw InvalidArgumentException("Quantization is supported only for FP32 tensors");
+ }
+
+ // Work out the range based on the input constants
+ unsigned int inputNumElements = input.GetNumElements();
+ const float* inputData = reinterpret_cast<const float*>(input.GetMemoryArea());
+
+ float min = std::numeric_limits<float>::max();
+ float max = std::numeric_limits<float>::lowest();
+
+ for (unsigned int i = 0; i < inputNumElements; i++)
+ {
+ const float inputValue = inputData[i];
+
+ min = std::min(min, inputValue);
+ max = std::max(max, inputValue);
+ }
+ SetRange(layer, 0, min, max);
+}
+
+void DynamicQuantizationVisitor::VisitMergerLayer(const IConnectableLayer* layer,
+ const OriginsDescriptor& mergerDescriptor,
+ const char* name)
+{
+ float min = std::numeric_limits<float>::max();
+ float max = std::numeric_limits<float>::lowest();
+ for (unsigned int i = 0; i < layer->GetNumInputSlots(); ++i)
+ {
+ const IOutputSlot* outputSlot = layer->GetInputSlot(i).GetConnection();
+ LayerGuid layerId = outputSlot->GetOwningLayerGuid();
+ unsigned int slotIndex = outputSlot->CalculateIndexOnOwner();
+ RangeTracker::MinMaxRange range = m_RangeTracker.GetRange(layerId, slotIndex);
+ min = std::min(min, range.first);
+ max = std::max(max, range.second);
+ }
+ SetRange(layer, 0, min, max);
+ AddToCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitReshapeLayer(const IConnectableLayer* layer,
+ const ReshapeDescriptor& reshapeDescriptor,
+ const char* name)
+{
+ boost::ignore_unused(reshapeDescriptor);
+ boost::ignore_unused(name);
+ AddToNonCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitSplitterLayer(const IConnectableLayer* layer,
+ const SplitterDescriptor& splitterDescriptor,
+ const char* name)
+{
+ boost::ignore_unused(splitterDescriptor);
+ boost::ignore_unused(name);
+ AddToNonCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitResizeBilinearLayer(const IConnectableLayer* layer,
+ const ResizeBilinearDescriptor& resizeDesc,
+ const char* name)
+{
+ boost::ignore_unused(resizeDesc);
+ boost::ignore_unused(name);
+ AddToNonCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitStridedSliceLayer(const IConnectableLayer* layer,
+ const StridedSliceDescriptor& stridedSliceDescriptor,
+ const char* name)
+{
+ boost::ignore_unused(stridedSliceDescriptor);
+ boost::ignore_unused(name);
+ AddToNonCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitBatchToSpaceNdLayer(const IConnectableLayer* layer,
+ const BatchToSpaceNdDescriptor& batchToSpaceNdDescriptor,
+ const char* name)
+{
+ boost::ignore_unused(batchToSpaceNdDescriptor);
+ boost::ignore_unused(name);
+ AddToNonCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitInputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name)
+{
+ boost::ignore_unused(id);
+ boost::ignore_unused(name);
+ SetRange(layer, 0, -0.0f, 0.0f);
+ AddToCalibratedLayers(layer);
+}
+
+void DynamicQuantizationVisitor::VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name)
+{
+ boost::ignore_unused(id);
+ boost::ignore_unused(name);
+ AddToNonCalibratedLayers(layer);
+ m_OutputLayers.push_back(id);
+}
+
+const std::vector<LayerBindingId>& DynamicQuantizationVisitor::GetOutputLayers()
+{
+ return m_OutputLayers;
+}
+
+} //namespace armnn