aboutsummaryrefslogtreecommitdiff
path: root/src/armnnTestUtils
diff options
context:
space:
mode:
authorTeresa Charlin <teresa.charlinreyes@arm.com>2023-04-03 19:57:00 +0100
committerColm Donelan <colm.donelan@arm.com>2023-04-18 17:27:41 +0000
commitacb3ec51e51542d3011ed87842f87c2261abaaff (patch)
treeb1ed73756c1db4a8e71b18a5a8256f42bb49341b /src/armnnTestUtils
parent8294e96a2f0f4ad3f5cd261079a6f90eee40142c (diff)
downloadarmnn-acb3ec51e51542d3011ed87842f87c2261abaaff.tar.gz
GitHub #719 Set quantization parameter scale to 1.0, instead of 0.0.
* Arm NN does not account for int8 or uint8 not quantized types, Tensorflow does. Not quantized int8 and uint8 is the same as quantized int8 and uint8 with scale = 1.0 and offset= 0 Default offset/zero_point was already 0, this review sets the default scale to 1.0. Signed-off-by: Teresa Charlin <teresa.charlinreyes@arm.com> Change-Id: Ibc3eecc281de516c2cc706e17bde01c64ff9556e
Diffstat (limited to 'src/armnnTestUtils')
-rw-r--r--src/armnnTestUtils/CreateWorkload.hpp32
1 files changed, 16 insertions, 16 deletions
diff --git a/src/armnnTestUtils/CreateWorkload.hpp b/src/armnnTestUtils/CreateWorkload.hpp
index 5e11ab6258..637f035365 100644
--- a/src/armnnTestUtils/CreateWorkload.hpp
+++ b/src/armnnTestUtils/CreateWorkload.hpp
@@ -521,8 +521,8 @@ std::unique_ptr<Convolution2dWorkload> CreateConvolution2dWorkloadTest(armnn::IW
layerDesc.m_BiasEnabled = false;
layerDesc.m_DataLayout = dataLayout;
- float inputsQScale = DataType == armnn::DataType::QAsymmU8 ? 1.0f : 0.0;
- float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 0.0;
+ float inputsQScale = 1.0f;
+ float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 1.0;
Convolution2dLayer* const layer = graph.AddLayer<Convolution2dLayer>(layerDesc, "layer");
@@ -585,8 +585,8 @@ std::unique_ptr<Convolution2dWorkload> CreateConvolution2dFusedActivationWithBlo
layerDesc.m_BiasEnabled = true;
layerDesc.m_DataLayout = dataLayout;
- float inputsQScale = DataType == armnn::DataType::QAsymmU8 ? 1.0f : 0.0;
- float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 0.0;
+ float inputsQScale = 1.0f;
+ float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 1.0;
Convolution2dLayer* const layer = graph.AddLayer<Convolution2dLayer>(layerDesc, "layer");
@@ -678,8 +678,8 @@ std::unique_ptr<Convolution2dWorkload> CreateConvolution2dWorkloadFastMathTest(a
layerDesc.m_BiasEnabled = true;
layerDesc.m_DataLayout = dataLayout;
- float inputsQScale = DataType == armnn::DataType::QAsymmU8 ? 1.0f : 0.0;
- float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 0.0;
+ float inputsQScale = 1.0f;
+ float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 1.0;
Convolution2dLayer* const layer = graph.AddLayer<Convolution2dLayer>(layerDesc, "layer");
@@ -1141,8 +1141,8 @@ std::unique_ptr<Convolution2dWorkload> CreateDirectConvolution2dWorkloadTest(arm
Convolution2dLayer* const layer = graph.AddLayer<Convolution2dLayer>(layerDesc, "layer");
- float inputsQScale = DataType == armnn::DataType::QAsymmU8 ? 1.0f : 0.0;
- float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 0.0;
+ float inputsQScale = 1.0f;
+ float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 1.0;
TensorShape biasShape = TensorShape{ 2 };
TensorShape weightShape = TensorShape{ 2, 3, 3, 3 };
@@ -1203,8 +1203,8 @@ std::unique_ptr<DepthwiseConvolution2dFloat32Workload> CreateDepthwiseConvolutio
layerDesc.m_BiasEnabled = false;
layerDesc.m_DataLayout = dataLayout;
- float inputsQScale = DataType == armnn::DataType::QAsymmU8 ? 1.0f : 0.0;
- float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 0.0;
+ float inputsQScale = 1.0f;
+ float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 1.0;
TensorShape weightShape({1, 4, 4, 2});
TensorShape inputShape = (dataLayout == DataLayout::NCHW) ?
@@ -1257,8 +1257,8 @@ std::unique_ptr<FullyConnectedWorkload> CreateFullyConnectedWorkloadTest(armnn::
FullyConnectedLayer* const layer = graph.AddLayer<FullyConnectedLayer>(layerDesc, "layer");
- float inputsQScale = DataType == armnn::DataType::QAsymmU8 ? 1.0f : 0.0;
- float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 0.0;
+ float inputsQScale = 1.0f;
+ float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 1.0;
armnn::TensorInfo weightsTensorInfo({7, 20}, DataType, inputsQScale);
weightsTensorInfo.SetConstant();
@@ -1302,8 +1302,8 @@ std::unique_ptr<FullyConnectedWorkload> CreateFullyConnectedWithBlobWorkloadTest
FullyConnectedLayer* const layer = graph.AddLayer<FullyConnectedLayer>(layerDesc, "layer");
- float inputsQScale = DataType == armnn::DataType::QAsymmU8 ? 1.0f : 0.0;
- float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 0.0;
+ float inputsQScale = 1.0f;
+ float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 1.0;
armnn::TensorInfo weightsTensorInfo({7, 20}, DataType, inputsQScale);
armnn::TensorInfo biasesTensorInfo({7}, GetBiasDataType(DataType), inputsQScale);
@@ -1378,8 +1378,8 @@ std::unique_ptr<FullyConnectedWorkload> CreateFullyConnectedWorkloadWeightsBiase
FullyConnectedLayer* const layer = graph.AddLayer<FullyConnectedLayer>(layerDesc, "layer");
- float inputsQScale = DataType == armnn::DataType::QAsymmU8 ? 1.0f : 0.0;
- float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 0.0;
+ float inputsQScale = 1.0f;
+ float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 1.0;
// Creates extra layers with weights and biases as input layers.
Layer* const input = graph.AddLayer<InputLayer>(1, "input");