From f9179d393a07eb9eed753e315df79d22391906c6 Mon Sep 17 00:00:00 2001 From: Michele Di Giorgio Date: Wed, 27 Nov 2019 16:17:30 +0000 Subject: COMPMID-2793: Add support for QASYMM8_SIGNED in CLGEMMLowpMatrixMultiplyReshapedOnlyRHSKernel Change-Id: I8abfdd3372cc394b98ec038b9fcb4abfe9216894 Signed-off-by: Michele Di Giorgio Reviewed-on: https://review.mlplatform.org/c/2401 Reviewed-by: Giorgio Arena Tested-by: Arm Jenkins Reviewed-by: Georgios Pinitas Comments-Addressed: Arm Jenkins --- tests/validation/fixtures/GEMMLowpFixture.h | 133 ++++++++++++++++++++-------- 1 file changed, 98 insertions(+), 35 deletions(-) (limited to 'tests/validation/fixtures/GEMMLowpFixture.h') diff --git a/tests/validation/fixtures/GEMMLowpFixture.h b/tests/validation/fixtures/GEMMLowpFixture.h index c17105edad..db52be5062 100644 --- a/tests/validation/fixtures/GEMMLowpFixture.h +++ b/tests/validation/fixtures/GEMMLowpFixture.h @@ -877,7 +877,8 @@ class GEMMLowpMatrixMultiplyReshapedOnlyRHSValidationFixture : public framework: { public: template - void setup(unsigned int m, unsigned int n, unsigned int k, unsigned int batch_size, unsigned int m0, unsigned int n0, unsigned int k0, unsigned int h0, bool interleave_rhs, bool transpose_rhs) + void setup(unsigned int m, unsigned int n, unsigned int k, unsigned int batch_size, unsigned int m0, unsigned int n0, + unsigned int k0, unsigned int h0, bool interleave_rhs, bool transpose_rhs, DataType data_type) { GEMMLHSMatrixInfo lhs_info; lhs_info.m0 = m0; @@ -894,24 +895,40 @@ public: const TensorShape lhs_shape(k, m, batch_size); const TensorShape rhs_shape(n, k, batch_size); - _target = compute_target(lhs_shape, rhs_shape, lhs_info, rhs_info); - _reference = compute_reference(lhs_shape, rhs_shape); + _target = compute_target(lhs_shape, rhs_shape, lhs_info, rhs_info, data_type); + _reference = compute_reference(lhs_shape, rhs_shape, data_type); } protected: template void fill(U &&tensor, int i) { - // Between 1 and 254 in order to avoid having -128 and 128 for the DOT product path - std::uniform_int_distribution<> distribution(1, 254); - library->fill(tensor, distribution, i); + switch(tensor.data_type()) + { + case DataType::QASYMM8: + { + // Between 1 and 254 in order to avoid having -128 and 128 for the DOT product path + std::uniform_int_distribution<> distribution(1, 254); + library->fill(tensor, distribution, i); + } + break; + case DataType::QASYMM8_SIGNED: + { + std::uniform_int_distribution<> distribution(-127, 126); + library->fill(tensor, distribution, i); + } + break; + default: + ARM_COMPUTE_ERROR("Unsupported data type"); + } } - TensorType compute_target(const TensorShape &lhs_shape, const TensorShape &rhs_shape, const GEMMLHSMatrixInfo &lhs_info, const GEMMRHSMatrixInfo &rhs_info) + TensorType compute_target(const TensorShape &lhs_shape, const TensorShape &rhs_shape, const GEMMLHSMatrixInfo &lhs_info, + const GEMMRHSMatrixInfo &rhs_info, DataType data_type) { // Create tensors - TensorType lhs = create_tensor(lhs_shape, DataType::QASYMM8, 1); - TensorType rhs = create_tensor(rhs_shape, DataType::QASYMM8, 1); + TensorType lhs = create_tensor(lhs_shape, data_type, 1); + TensorType rhs = create_tensor(rhs_shape, data_type, 1); TensorType rhs_reshaped; TensorType dst; @@ -952,21 +969,36 @@ protected: return dst; } - SimpleTensor compute_reference(const TensorShape &lhs_shape, const TensorShape &rhs_shape) + SimpleTensor compute_reference(const TensorShape &lhs_shape, const TensorShape &rhs_shape, DataType data_type) { TensorShape dst_shape = lhs_shape; dst_shape[0] = rhs_shape[0]; dst_shape[1] = lhs_shape[1]; - // Create reference - SimpleTensor lhs{ lhs_shape, DataType::QASYMM8, 1 }; - SimpleTensor rhs{ rhs_shape, DataType::QASYMM8, 1 }; + if(data_type == DataType::QASYMM8) + { + // Create reference + SimpleTensor lhs{ lhs_shape, data_type, 1 }; + SimpleTensor rhs{ rhs_shape, data_type, 1 }; - // Fill reference - fill(lhs, 0); - fill(rhs, 1); + // Fill reference + fill(lhs, 0); + fill(rhs, 1); - return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + } + else + { + // Create reference + SimpleTensor lhs{ lhs_shape, data_type, 1 }; + SimpleTensor rhs{ rhs_shape, data_type, 1 }; + + // Fill reference + fill(lhs, 0); + fill(rhs, 1); + + return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + } } TensorType _target{}; @@ -978,8 +1010,8 @@ class GEMMLowpMatrixMultiplyReshapedOnlyRHS3DValidationFixture : public framewor { public: template - void setup(unsigned int m_w, unsigned int m_h, unsigned int n, unsigned int k, unsigned int batch_size, unsigned int m0, unsigned int n0, unsigned int k0, unsigned int h0, - bool interleave_rhs, bool transpose_rhs) + void setup(unsigned int m_w, unsigned int m_h, unsigned int n, unsigned int k, unsigned int batch_size, unsigned int m0, unsigned int n0, + unsigned int k0, unsigned int h0, bool interleave_rhs, bool transpose_rhs, DataType data_type) { GEMMLHSMatrixInfo lhs_info; lhs_info.m0 = m0; @@ -999,24 +1031,40 @@ public: const TensorShape lhs_shape(k, m, batch_size); const TensorShape rhs_shape(n, k, batch_size); - _target = compute_target(lhs_shape, rhs_shape, lhs_info, rhs_info, m_h); - _reference = compute_reference(lhs_shape, rhs_shape, m_h); + _target = compute_target(lhs_shape, rhs_shape, lhs_info, rhs_info, m_h, data_type); + _reference = compute_reference(lhs_shape, rhs_shape, m_h, data_type); } protected: template void fill(U &&tensor, int i) { - // Between 1 and 254 in order to avoid having -128 and 128 for the DOT product path - std::uniform_int_distribution<> distribution(1, 254); - library->fill(tensor, distribution, i); + switch(tensor.data_type()) + { + case DataType::QASYMM8: + { + // Between 1 and 254 in order to avoid having -128 and 128 for the DOT product path + std::uniform_int_distribution<> distribution(1, 254); + library->fill(tensor, distribution, i); + } + break; + case DataType::QASYMM8_SIGNED: + { + std::uniform_int_distribution<> distribution(-127, 126); + library->fill(tensor, distribution, i); + } + break; + default: + ARM_COMPUTE_ERROR("Unsupported data type"); + } } - TensorType compute_target(const TensorShape &lhs_shape, const TensorShape &rhs_shape, const GEMMLHSMatrixInfo &lhs_info, const GEMMRHSMatrixInfo &rhs_info, unsigned int m_h) + TensorType compute_target(const TensorShape &lhs_shape, const TensorShape &rhs_shape, const GEMMLHSMatrixInfo &lhs_info, + const GEMMRHSMatrixInfo &rhs_info, unsigned int m_h, DataType data_type) { // Create tensors - TensorType lhs = create_tensor(lhs_shape, DataType::QASYMM8, 1); - TensorType rhs = create_tensor(rhs_shape, DataType::QASYMM8, 1); + TensorType lhs = create_tensor(lhs_shape, data_type, 1); + TensorType rhs = create_tensor(rhs_shape, data_type, 1); TensorType rhs_reshaped; TensorType dst; @@ -1057,7 +1105,7 @@ protected: return dst; } - SimpleTensor compute_reference(const TensorShape &lhs_shape, const TensorShape &rhs_shape, unsigned int m_h) + SimpleTensor compute_reference(const TensorShape &lhs_shape, const TensorShape &rhs_shape, unsigned int m_h, DataType data_type) { TensorShape dst_shape = lhs_shape; dst_shape.set(0, rhs_shape[0]); @@ -1065,15 +1113,30 @@ protected: dst_shape.set(2, m_h); dst_shape.set(3, lhs_shape[2]); - // Create reference - SimpleTensor lhs{ lhs_shape, DataType::QASYMM8, 1 }; - SimpleTensor rhs{ rhs_shape, DataType::QASYMM8, 1 }; + if(data_type == DataType::QASYMM8) + { + // Create reference + SimpleTensor lhs{ lhs_shape, data_type, 1 }; + SimpleTensor rhs{ rhs_shape, data_type, 1 }; - // Fill reference - fill(lhs, 0); - fill(rhs, 1); + // Fill reference + fill(lhs, 0); + fill(rhs, 1); - return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + } + else + { + // Create reference + SimpleTensor lhs{ lhs_shape, data_type, 1 }; + SimpleTensor rhs{ rhs_shape, data_type, 1 }; + + // Fill reference + fill(lhs, 0); + fill(rhs, 1); + + return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + } } TensorType _target{}; -- cgit v1.2.1