From 598950d611304ffeb9d57a5b28a13b0ddd629026 Mon Sep 17 00:00:00 2001 From: FinnWilliamsArm Date: Thu, 18 Jul 2019 11:15:42 +0100 Subject: IVGCVSW-3453 Fix VTS quant_output_multiplier_gt_1 test failures * Remove ValidateTensorQuantizationMultiplier * Update CL pin to COMPMID-2336 Signed-off-by: FinnWilliamsArm Change-Id: Iaece5b564134f7ec91f82cb246bcb10ad455999d --- src/backends/backendsCommon/WorkloadData.cpp | 35 ---------------------------- 1 file changed, 35 deletions(-) (limited to 'src/backends') diff --git a/src/backends/backendsCommon/WorkloadData.cpp b/src/backends/backendsCommon/WorkloadData.cpp index 878602391c..455e675fef 100644 --- a/src/backends/backendsCommon/WorkloadData.cpp +++ b/src/backends/backendsCommon/WorkloadData.cpp @@ -262,27 +262,6 @@ void ValidateBroadcastTensorShapesMatch(const TensorInfo& first, } } -//--------------------------------------------------------------- -/// Validates that the output tensor's quantization scale is greater than the product -/// of the two input tensors' quantization scales. This is a requirement of the implementation of -/// the quantized multiplication. -void ValidateTensorQuantizationMultiplier(const TensorInfo& inputTensor1, const TensorInfo& inputTensor2, - const TensorInfo& outputTensorInfo, std::string const& descName, - const std::string& inputTensor1Name, const std::string& inputTensor2Name, const std::string& outputTensorName) -{ - if (outputTensorInfo.GetDataType() == DataType::QuantisedAsymm8) - { - if (outputTensorInfo.GetQuantizationScale() <= - inputTensor1.GetQuantizationScale() * inputTensor2.GetQuantizationScale()) - { - std::stringstream msg; - msg << descName << ": Quantization scale of " << outputTensorName << " is not greater than " << - "the product of the " << inputTensor1Name << " and " << inputTensor2Name << " tensors"; - throw InvalidArgumentException(msg.str()); - } - } -} - //--------------------------------------------------------------- void ValidateDataTypes(const TensorInfo& info, const std::vector& supportedTypes, @@ -705,8 +684,6 @@ void FullyConnectedQueueDescriptor::Validate(const WorkloadInfo& workloadInfo) c ValidateTensorNumDimensions(m_Bias->GetTensorInfo(), "FullyConnectedQueueDescriptor", 1, "bias"); } - ValidateTensorQuantizationMultiplier(workloadInfo.m_InputTensorInfos[0], m_Weight->GetTensorInfo(), - workloadInfo.m_OutputTensorInfos[0], "FullyConnectedQueueDescriptor", "input", "weights", "output"); // Check the supported data types std::vector supportedTypes = @@ -892,8 +869,6 @@ void Convolution2dQueueDescriptor::Validate(const WorkloadInfo& workloadInfo) co workloadInfo.m_InputTensorInfos[0], m_Weight->GetTensorInfo(), "Convolution2dQueueDescriptor"); } - ValidateTensorQuantizationMultiplier(workloadInfo.m_InputTensorInfos[0], m_Weight->GetTensorInfo(), - workloadInfo.m_OutputTensorInfos[0], "Convolution2dQueueDescriptor", "input", "weights", "output"); } void DepthwiseConvolution2dQueueDescriptor::Validate(const WorkloadInfo& workloadInfo) const @@ -945,9 +920,6 @@ void DepthwiseConvolution2dQueueDescriptor::Validate(const WorkloadInfo& workloa "DepthwiseConvolution2dQueueDescriptor", "bias"); } - ValidateTensorQuantizationMultiplier(workloadInfo.m_InputTensorInfos[0], m_Weight->GetTensorInfo(), - workloadInfo.m_OutputTensorInfos[0], "DepthwiseConvolution2dQueueDescriptor", "input", "weights", "output"); - // Check the supported data types std::vector supportedTypes = { DataType::Float32, @@ -2256,13 +2228,6 @@ void TransposeConvolution2dQueueDescriptor::Validate(const WorkloadInfo& workloa descriptorName); } - ValidateTensorQuantizationMultiplier(workloadInfo.m_InputTensorInfos[0], - m_Weight->GetTensorInfo(), - workloadInfo.m_OutputTensorInfos[0], - descriptorName, - "input", - "weights", - "output"); } } //namespace armnn -- cgit v1.2.1