From 28287afbea9549e8e2904084ae895c04cca88e95 Mon Sep 17 00:00:00 2001 From: Sheri Zhang Date: Tue, 25 Feb 2020 14:13:54 +0000 Subject: COMPMID-2792: Add support for QASYMM8_SIGNED in CLGEMMLowpMatrixMultiplyReshapedKernel Signed-off-by: Sheri Zhang Change-Id: I005e604253394f31173f37ec0296caf76b5e697c Reviewed-on: https://eu-gerrit-1.euhpc.arm.com/c/VisualCompute/ComputeLibrary/+/227008 Tested-by: bsgcomp Reviewed-by: Sang-Hoon Park Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/2798 Tested-by: Arm Jenkins Comments-Addressed: Arm Jenkins Reviewed-by: Michele Di Giorgio --- .../CL/GEMMLowpMatrixMultiplyReshaped.cpp | 78 ++++++++++-- tests/validation/fixtures/GEMMLowpFixture.h | 139 ++++++++++++++++----- 2 files changed, 170 insertions(+), 47 deletions(-) (limited to 'tests') diff --git a/tests/validation/CL/GEMMLowpMatrixMultiplyReshaped.cpp b/tests/validation/CL/GEMMLowpMatrixMultiplyReshaped.cpp index e1f2b0c9a9..96e41b3f54 100644 --- a/tests/validation/CL/GEMMLowpMatrixMultiplyReshaped.cpp +++ b/tests/validation/CL/GEMMLowpMatrixMultiplyReshaped.cpp @@ -1,5 +1,5 @@ /* - * Copyright (c) 2019 ARM Limited. + * Copyright (c) 2019-2020 ARM Limited. * * SPDX-License-Identifier: MIT * @@ -81,7 +81,8 @@ const auto k_values = framework::dataset::make("K", 23); const auto b_values = framework::dataset::make("batch_size", 1, 3); /** M0 values to test - Precommit */ -const auto m0_values_precommit = framework::dataset::make("M0", {4, 6}); +const auto m0_values_precommit_1 = framework::dataset::make("M0", { 4 }); +const auto m0_values_precommit_2 = framework::dataset::make("M0", { 6 }); /** N0 values to test - Precommit */ const auto n0_values_precommit = framework::dataset::make("N0", { 4 }); @@ -119,26 +120,31 @@ const auto i_values_rhs = framework::dataset::make("interleave_rhs", { true, fal TEST_SUITE(CL) TEST_SUITE(GEMMLowpMatrixMultiplyReshaped) + +TEST_SUITE(QUANTIZED) + +TEST_SUITE(QASYMM8) FIXTURE_DATA_TEST_CASE(RunSmall, CLGEMMLowpMatrixMultiplyReshapedFixture, framework::DatasetMode::ALL, - combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( + combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( m_values, n_values), k_values), b_values), - m0_values_precommit), + m0_values_precommit_1), n0_values_precommit), k0_values_precommit), v0_values_precommit), h0_values_precommit), i_values_lhs), - i_values_rhs)) + i_values_rhs), + framework::dataset::make("DataType", { DataType::QASYMM8 }))) { // Validate output validate(CLAccessor(_target), _reference); } FIXTURE_DATA_TEST_CASE(RunLarge, CLGEMMLowpMatrixMultiplyReshapedFixture, framework::DatasetMode::DISABLED, - combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( + combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( m_values, n_values), k_values), @@ -149,33 +155,35 @@ FIXTURE_DATA_TEST_CASE(RunLarge, CLGEMMLowpMatrixMultiplyReshapedFixture, framew v0_values_nightly), h0_values_nightly), i_values_lhs), - i_values_rhs)) + i_values_rhs), + framework::dataset::make("DataType", { DataType::QASYMM8 }))) { // Validate output validate(CLAccessor(_target), _reference); } FIXTURE_DATA_TEST_CASE(RunSmall3D, CLGEMMLowpMatrixMultiplyReshaped3DFixture, framework::DatasetMode::ALL, - combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( + combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( m_w_values, m_h_values), n_values), k_values), b_values), - m0_values_precommit), + m0_values_precommit_1), n0_values_precommit), k0_values_precommit), v0_values_precommit), h0_values_precommit), i_values_lhs), - i_values_rhs)) + i_values_rhs), + framework::dataset::make("DataType", { DataType::QASYMM8 }))) { // Validate output validate(CLAccessor(_target), _reference); } FIXTURE_DATA_TEST_CASE(RunLarge3D, CLGEMMLowpMatrixMultiplyReshaped3DFixture, framework::DatasetMode::DISABLED, - combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( + combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( m_w_values, m_h_values), n_values), @@ -187,13 +195,57 @@ FIXTURE_DATA_TEST_CASE(RunLarge3D, CLGEMMLowpMatrixMultiplyReshaped3DFixture, fr v0_values_nightly), h0_values_nightly), i_values_lhs), - i_values_rhs)) + i_values_rhs), + framework::dataset::make("DataType", { DataType::QASYMM8 }))) +{ + // Validate output + validate(CLAccessor(_target), _reference); +} +TEST_SUITE_END() // QASYMM8 + +TEST_SUITE(QASYMM8_SIGNED) +FIXTURE_DATA_TEST_CASE(RunSmall, CLGEMMLowpMatrixMultiplyReshapedFixture, framework::DatasetMode::ALL, + combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( + m_values, + n_values), + k_values), + b_values), + m0_values_precommit_2), + n0_values_precommit), + k0_values_precommit), + v0_values_precommit), + h0_values_precommit), + i_values_lhs), + i_values_rhs), + framework::dataset::make("DataType", { DataType::QASYMM8_SIGNED }))) { // Validate output validate(CLAccessor(_target), _reference); } +FIXTURE_DATA_TEST_CASE(RunSmall3D, CLGEMMLowpMatrixMultiplyReshaped3DFixture, framework::DatasetMode::ALL, + combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine(combine( + m_w_values, + m_h_values), + n_values), + k_values), + b_values), + m0_values_precommit_2), + n0_values_precommit), + k0_values_precommit), + v0_values_precommit), + h0_values_precommit), + i_values_lhs), + i_values_rhs), + framework::dataset::make("DataType", { DataType::QASYMM8_SIGNED }))) +{ + // Validate output + validate(CLAccessor(_target), _reference); +} +TEST_SUITE_END() // QASYMM8_SIGNED + +TEST_SUITE_END() // QUANTIZED TEST_SUITE_END() // GEMMLowpMatrixMultiplyReshaped TEST_SUITE_END() // CL } // namespace validation } // namespace test -} // namespace arm_compute \ No newline at end of file +} // namespace arm_compute diff --git a/tests/validation/fixtures/GEMMLowpFixture.h b/tests/validation/fixtures/GEMMLowpFixture.h index 1154d6c8de..d13ada9d64 100644 --- a/tests/validation/fixtures/GEMMLowpFixture.h +++ b/tests/validation/fixtures/GEMMLowpFixture.h @@ -656,7 +656,7 @@ class GEMMLowpMatrixMultiplyReshapedValidationFixture : public framework::Fixtur public: template void setup(unsigned int m, unsigned int n, unsigned int k, unsigned int batch_size, unsigned int m0, unsigned int n0, unsigned int k0, unsigned int v0, unsigned int h0, bool interleave_lhs, - bool interleave_rhs) + bool interleave_rhs, DataType data_type) { GEMMLHSMatrixInfo lhs_info; lhs_info.m0 = m0; @@ -676,24 +676,39 @@ public: const TensorShape lhs_shape(k, m, batch_size); const TensorShape rhs_shape(n, k, batch_size); - _target = compute_target(lhs_shape, rhs_shape, lhs_info, rhs_info); - _reference = compute_reference(lhs_shape, rhs_shape); + _target = compute_target(lhs_shape, rhs_shape, lhs_info, rhs_info, data_type); + _reference = compute_reference(lhs_shape, rhs_shape, data_type); } protected: template void fill(U &&tensor, int i) { - // Between 1 and 254 in order to avoid having -128 and 128 for the DOT product path - std::uniform_int_distribution<> distribution(1, 254); - library->fill(tensor, distribution, i); + switch(tensor.data_type()) + { + case DataType::QASYMM8: + { + // Between 1 and 254 in order to avoid having -128 and 128 for the DOT product path + std::uniform_int_distribution<> distribution(1, 254); + library->fill(tensor, distribution, i); + } + break; + case DataType::QASYMM8_SIGNED: + { + std::uniform_int_distribution<> distribution(-127, 126); + library->fill(tensor, distribution, i); + } + break; + default: + ARM_COMPUTE_ERROR("Unsupported data type"); + } } - TensorType compute_target(const TensorShape &lhs_shape, const TensorShape &rhs_shape, const GEMMLHSMatrixInfo &lhs_info, const GEMMRHSMatrixInfo &rhs_info) + TensorType compute_target(const TensorShape &lhs_shape, const TensorShape &rhs_shape, const GEMMLHSMatrixInfo &lhs_info, const GEMMRHSMatrixInfo &rhs_info, DataType data_type) { // Create tensors - TensorType lhs = create_tensor(lhs_shape, DataType::QASYMM8, 1); - TensorType rhs = create_tensor(rhs_shape, DataType::QASYMM8, 1); + TensorType lhs = create_tensor(lhs_shape, data_type, 1); + TensorType rhs = create_tensor(rhs_shape, data_type, 1); TensorType lhs_reshaped; TensorType rhs_reshaped; TensorType dst; @@ -740,21 +755,41 @@ protected: return dst; } - SimpleTensor compute_reference(const TensorShape &lhs_shape, const TensorShape &rhs_shape) + SimpleTensor compute_reference(const TensorShape &lhs_shape, const TensorShape &rhs_shape, DataType data_type) { TensorShape dst_shape = lhs_shape; dst_shape[0] = rhs_shape[0]; dst_shape[1] = lhs_shape[1]; - // Create reference - SimpleTensor lhs{ lhs_shape, DataType::QASYMM8, 1 }; - SimpleTensor rhs{ rhs_shape, DataType::QASYMM8, 1 }; + switch(data_type) + { + case DataType::QASYMM8: + { + // Create reference + SimpleTensor lhs{ lhs_shape, data_type, 1 }; + SimpleTensor rhs{ rhs_shape, data_type, 1 }; - // Fill reference - fill(lhs, 0); - fill(rhs, 1); + // Fill reference + fill(lhs, 0); + fill(rhs, 1); - return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + } + case DataType::QASYMM8_SIGNED: + { + // Create reference + SimpleTensor lhs{ lhs_shape, data_type, 1 }; + SimpleTensor rhs{ rhs_shape, data_type, 1 }; + + // Fill reference + fill(lhs, 0); + fill(rhs, 1); + + return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + } + default: + ARM_COMPUTE_ERROR("Unsupported data type"); + } } TensorType _target{}; @@ -767,7 +802,7 @@ class GEMMLowpMatrixMultiplyReshaped3DValidationFixture : public framework::Fixt public: template void setup(unsigned int m_w, unsigned int m_h, unsigned int n, unsigned int k, unsigned int batch_size, unsigned int m0, unsigned int n0, unsigned int k0, unsigned int v0, unsigned int h0, - bool interleave_lhs, bool interleave_rhs) + bool interleave_lhs, bool interleave_rhs, DataType data_type) { GEMMLHSMatrixInfo lhs_info; lhs_info.m0 = m0; @@ -790,24 +825,40 @@ public: const TensorShape lhs_shape(k, m, batch_size); const TensorShape rhs_shape(n, k, batch_size); - _target = compute_target(lhs_shape, rhs_shape, lhs_info, rhs_info, m_h); - _reference = compute_reference(lhs_shape, rhs_shape, m_h); + _target = compute_target(lhs_shape, rhs_shape, lhs_info, rhs_info, m_h, data_type); + _reference = compute_reference(lhs_shape, rhs_shape, m_h, data_type); } protected: template void fill(U &&tensor, int i) { - // Between 1 and 254 in order to avoid having -128 and 128 for the DOT product path - std::uniform_int_distribution<> distribution(1, 254); - library->fill(tensor, distribution, i); + switch(tensor.data_type()) + { + case DataType::QASYMM8: + { + // Between 1 and 254 in order to avoid having -128 and 128 for the DOT product path + std::uniform_int_distribution<> distribution(1, 254); + library->fill(tensor, distribution, i); + } + break; + case DataType::QASYMM8_SIGNED: + { + std::uniform_int_distribution<> distribution(-127, 126); + library->fill(tensor, distribution, i); + } + break; + default: + ARM_COMPUTE_ERROR("Unsupported data type"); + } } - TensorType compute_target(const TensorShape &lhs_shape, const TensorShape &rhs_shape, const GEMMLHSMatrixInfo &lhs_info, const GEMMRHSMatrixInfo &rhs_info, unsigned int m_h) + TensorType compute_target(const TensorShape &lhs_shape, const TensorShape &rhs_shape, const GEMMLHSMatrixInfo &lhs_info, const GEMMRHSMatrixInfo &rhs_info, unsigned int m_h, + DataType data_type) { // Create tensors - TensorType lhs = create_tensor(lhs_shape, DataType::QASYMM8, 1); - TensorType rhs = create_tensor(rhs_shape, DataType::QASYMM8, 1); + TensorType lhs = create_tensor(lhs_shape, data_type, 1); + TensorType rhs = create_tensor(rhs_shape, data_type, 1); TensorType lhs_reshaped; TensorType rhs_reshaped; TensorType dst; @@ -854,7 +905,7 @@ protected: return dst; } - SimpleTensor compute_reference(const TensorShape &lhs_shape, const TensorShape &rhs_shape, unsigned int m_h) + SimpleTensor compute_reference(const TensorShape &lhs_shape, const TensorShape &rhs_shape, unsigned int m_h, DataType data_type) { TensorShape dst_shape = lhs_shape; dst_shape.set(0, rhs_shape[0]); @@ -862,15 +913,35 @@ protected: dst_shape.set(2, m_h); dst_shape.set(3, lhs_shape[2]); - // Create reference - SimpleTensor lhs{ lhs_shape, DataType::QASYMM8, 1 }; - SimpleTensor rhs{ rhs_shape, DataType::QASYMM8, 1 }; + switch(data_type) + { + case DataType::QASYMM8: + { + // Create reference + SimpleTensor lhs{ lhs_shape, data_type, 1 }; + SimpleTensor rhs{ rhs_shape, data_type, 1 }; - // Fill reference - fill(lhs, 0); - fill(rhs, 1); + // Fill reference + fill(lhs, 0); + fill(rhs, 1); - return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + } + case DataType::QASYMM8_SIGNED: + { + // Create reference + SimpleTensor lhs{ lhs_shape, data_type, 1 }; + SimpleTensor rhs{ rhs_shape, data_type, 1 }; + + // Fill reference + fill(lhs, 0); + fill(rhs, 1); + + return reference::gemmlowp_matrix_multiply_core(lhs, rhs, dst_shape, 0, 0); + } + default: + ARM_COMPUTE_ERROR("Unsupported data type"); + } } TensorType _target{}; -- cgit v1.2.1