From 0c2eeac6347533a1d3d456aebea492f5123388f3 Mon Sep 17 00:00:00 2001 From: Keith Davis Date: Tue, 11 Feb 2020 16:51:50 +0000 Subject: IVGCVSW-4436 Add ExecuteNetwork test for mobilenet_v2_int8 * Add QAsymmS8 to QueueDescriptor supportedTypes * Add QSymmS8/QAsymmS8 to RefLayerSupport supportedTypes * Some additional comments and refactoring Change-Id: I8567314452e6e8f6f69cb6e458ee147d3fc92fab Signed-off-by: Keith Davis --- src/backends/reference/RefLayerSupport.cpp | 43 +++++++++++++--------- src/backends/reference/RefWorkloadFactory.cpp | 25 ++++++++----- .../reference/workloads/RefDebugWorkload.hpp | 12 +++--- 3 files changed, 47 insertions(+), 33 deletions(-) (limited to 'src/backends/reference') diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp index c60348e529..bba83e23d4 100644 --- a/src/backends/reference/RefLayerSupport.cpp +++ b/src/backends/reference/RefLayerSupport.cpp @@ -4,15 +4,11 @@ // #include "RefLayerSupport.hpp" -#include "RefBackendId.hpp" +#include #include #include -#include -#include - -#include #include #include @@ -21,7 +17,6 @@ #include #include -#include #include using namespace boost; @@ -84,9 +79,11 @@ bool RefLayerSupport::IsActivationSupported(const TensorInfo& input, bool supported = true; // Define supported types. - std::array supportedTypes = { + std::array supportedTypes = { DataType::Float32, DataType::Float16, + DataType::QSymmS8, + DataType::QAsymmS8, DataType::QAsymmU8, DataType::QSymmS16 }; @@ -147,10 +144,11 @@ bool RefLayerSupport::IsAdditionSupported(const TensorInfo& input0, { bool supported = true; - std::array supportedTypes = { + std::array supportedTypes = { DataType::Float32, DataType::Float16, DataType::QSymmS8, + DataType::QAsymmS8, DataType::QAsymmU8, DataType::QSymmS16 }; @@ -420,11 +418,12 @@ bool RefLayerSupport::IsConvolution2dSupported(const TensorInfo& input, bool supported = true; // Define supported types. - std::array supportedTypes = + std::array supportedTypes = { DataType::Float32, DataType::Float16, DataType::QAsymmU8, + DataType::QAsymmS8, DataType::QSymmS8, DataType::QSymmS16 }; @@ -439,13 +438,14 @@ bool RefLayerSupport::IsConvolution2dSupported(const TensorInfo& input, "Reference Convolution2d: input and output types mismatched."); const DataType inputType = input.GetDataType(); - if (inputType == DataType::QAsymmU8) + if (IsQuantized8BitType(inputType)) { ARMNN_NO_DEPRECATE_WARN_BEGIN - std::array supportedWeightTypes = + std::array supportedWeightTypes = { DataType::QAsymmU8, DataType::QSymmS8, + DataType::QAsymmS8, DataType::QuantizedSymm8PerAxis // deprecated }; ARMNN_NO_DEPRECATE_WARN_END @@ -485,11 +485,12 @@ bool RefLayerSupport::IsDebugSupported(const TensorInfo& input, { bool supported = true; - std::array supportedTypes = + std::array supportedTypes = { DataType::Float16, DataType::Float32, DataType::QAsymmU8, + DataType::QAsymmS8, DataType::QSymmS8, DataType::QSymmS16, DataType::Signed32 @@ -545,10 +546,12 @@ bool RefLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input, bool supported = true; // Define supported types. - std::array supportedTypes = + std::array supportedTypes = { DataType::Float32, DataType::Float16, + DataType::QSymmS8, + DataType::QAsymmS8, DataType::QAsymmU8, DataType::QSymmS16 }; @@ -572,7 +575,7 @@ bool RefLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input, ARMNN_NO_DEPRECATE_WARN_END const DataType inputType = input.GetDataType(); - if (inputType == DataType::QAsymmU8) + if (IsQuantized8BitType(inputType)) { supported &= CheckSupportRule(TypeAnyOf(weights, supportedWeightTypes), reasonIfUnsupported, @@ -1413,10 +1416,12 @@ bool RefLayerSupport::IsPooling2dSupported(const TensorInfo& input, bool supported = true; // Define supported output and inputs types. - std::array supportedTypes = + std::array supportedTypes = { DataType::Float32, DataType::Float16, + DataType::QSymmS8, + DataType::QAsymmS8, DataType::QAsymmU8, DataType::QSymmS16 }; @@ -1476,15 +1481,17 @@ bool RefLayerSupport::IsReshapeSupported(const TensorInfo& input, ignore_unused(output); ignore_unused(descriptor); // Define supported output types. - std::array supportedOutputTypes = + std::array supportedOutputTypes = { DataType::Float32, DataType::Float16, DataType::Signed32, + DataType::QAsymmS8, DataType::QAsymmU8, DataType::QSymmS8, DataType::QSymmS16 }; + return CheckSupportRule(TypeAnyOf(input, supportedOutputTypes), reasonIfUnsupported, "Reference reshape: input type not supported."); } @@ -1586,10 +1593,12 @@ bool RefLayerSupport::IsSoftmaxSupported(const TensorInfo& input, { boost::ignore_unused(descriptor); bool supported = true; - std::array supportedTypes = + std::array supportedTypes = { DataType::Float32, DataType::Float16, + DataType::QSymmS8, + DataType::QAsymmS8, DataType::QAsymmU8, DataType::QSymmS16 }; diff --git a/src/backends/reference/RefWorkloadFactory.cpp b/src/backends/reference/RefWorkloadFactory.cpp index 792bd7d3ad..dadb456104 100644 --- a/src/backends/reference/RefWorkloadFactory.cpp +++ b/src/backends/reference/RefWorkloadFactory.cpp @@ -45,17 +45,22 @@ bool IsDataType(const WorkloadInfo& info) return false; } +bool IsSigned32(const WorkloadInfo& info) +{ + return IsDataType(info); +} + bool IsFloat16(const WorkloadInfo& info) { return IsDataType(info); } -bool IsQSymm16(const WorkloadInfo& info) +bool IsQSymmS16(const WorkloadInfo& info) { return IsDataType(info); } -bool IsQSymm8(const WorkloadInfo& info) +bool IsQSymmS8(const WorkloadInfo& info) { return IsDataType(info); } @@ -187,20 +192,20 @@ std::unique_ptr RefWorkloadFactory::CreateDebug(const DebugQueueDescr { return std::make_unique(descriptor, info); } - if (IsQSymm16(info)) + if (IsQSymmS16(info)) { - return std::make_unique(descriptor, info); + return std::make_unique(descriptor, info); } - if (IsQSymm8(info)) + if (IsQSymmS8(info)) { - return std::make_unique(descriptor, info); + return std::make_unique(descriptor, info); } - if (IsDataType(info)) + if (IsSigned32(info)) { return std::make_unique(descriptor, info); } - return MakeWorkload(descriptor, info); + return MakeWorkload(descriptor, info); } std::unique_ptr RefWorkloadFactory::CreateDepthToSpace(const DepthToSpaceQueueDescriptor& descriptor, @@ -410,7 +415,7 @@ std::unique_ptr RefWorkloadFactory::CreateOutput(const OutputQueueDes std::unique_ptr RefWorkloadFactory::CreatePad(const PadQueueDescriptor& descriptor, const WorkloadInfo& info) const { - if (IsQSymm16(info)) + if (IsQSymmS16(info)) { return std::make_unique(descriptor, info); } @@ -424,7 +429,7 @@ std::unique_ptr RefWorkloadFactory::CreatePad(const PadQueueDescripto std::unique_ptr RefWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& descriptor, const WorkloadInfo& info) const { - if (IsQSymm16(info)) + if (IsQSymmS16(info)) { return std::make_unique(descriptor, info); } diff --git a/src/backends/reference/workloads/RefDebugWorkload.hpp b/src/backends/reference/workloads/RefDebugWorkload.hpp index a15a863892..4966ca3432 100644 --- a/src/backends/reference/workloads/RefDebugWorkload.hpp +++ b/src/backends/reference/workloads/RefDebugWorkload.hpp @@ -37,11 +37,11 @@ private: DebugCallbackFunction m_Callback; }; -using RefDebugFloat16Workload = RefDebugWorkload; -using RefDebugFloat32Workload = RefDebugWorkload; -using RefDebugQAsymm8Workload = RefDebugWorkload; -using RefDebugQSymm16Workload = RefDebugWorkload; -using RefDebugQSymm8Workload = RefDebugWorkload; -using RefDebugSigned32Workload = RefDebugWorkload; +using RefDebugFloat16Workload = RefDebugWorkload; +using RefDebugFloat32Workload = RefDebugWorkload; +using RefDebugQAsymmU8Workload = RefDebugWorkload; +using RefDebugQSymmS16Workload = RefDebugWorkload; +using RefDebugQSymmS8Workload = RefDebugWorkload; +using RefDebugSigned32Workload = RefDebugWorkload; } // namespace armnn -- cgit v1.2.1