From 215b4ea6c9dee480a22070d5873b0b8cb52531a0 Mon Sep 17 00:00:00 2001 From: Gian Marco Iodice Date: Thu, 28 Jun 2018 16:29:29 +0100 Subject: COMPMID-1277 - Optimizing CLIm2ColKernel for NHWC. This patch includes: - Im2Col optimizations for NHWC using a new data layout - Refactoring of CLIm2ColKernel adding validation method and auto-init - Removed im2col_reduced from CLIm2ColKernel and created a new kernel CLFlattenLayerKernel Change-Id: I1620640b6796baa268324b33ae92cdd8de53e27c Reviewed-on: https://eu-gerrit-1.euhpc.arm.com/141241 Tested-by: Jenkins Reviewed-by: Giorgio Arena --- tests/validation/CL/Im2Col.cpp | 54 +++++++++++++++++++------------- tests/validation/CL/LocallyConnected.cpp | 17 ++++++---- 2 files changed, 44 insertions(+), 27 deletions(-) (limited to 'tests/validation/CL') diff --git a/tests/validation/CL/Im2Col.cpp b/tests/validation/CL/Im2Col.cpp index 9422fcc49b..291befa6cb 100644 --- a/tests/validation/CL/Im2Col.cpp +++ b/tests/validation/CL/Im2Col.cpp @@ -41,8 +41,18 @@ namespace validation { namespace { -const auto conv_filter_sizes = framework::dataset::make("KernelDims", { Size2D(3U, 3U), Size2D(3U, 1U), Size2D(1U, 5U), Size2D(5U, 5U), Size2D(7U, 7U) }); -const auto padstrides = framework::dataset::make("PadStride", { PadStrideInfo(1U, 1U, 0U, 0U), PadStrideInfo(1U, 1U, 1U, 1U), PadStrideInfo(2U, 2U, 0U, 2U) }); +// *INDENT-OFF* +// clang-format off +const auto conv_filter_sizes = framework::dataset::make("KernelDims", { Size2D(3U, 3U), + Size2D(5U, 5U), + Size2D(3U, 1U), + Size2D(1U, 3U), + Size2D(5U, 3U), + Size2D(1U, 1U), + Size2D(11U, 11U)} ); +const auto padstrides = framework::dataset::make("PadStride", { PadStrideInfo(1U, 1U, 0U, 0U), + PadStrideInfo(1U, 1U, 1U, 1U), + PadStrideInfo(2U, 2U, 0U, 2U) }); const auto conv_args = combine(combine(combine(conv_filter_sizes, padstrides), framework::dataset::make("QuantizationInfo", QuantizationInfo(0.5f, 10))), framework::dataset::make("DataLayout", { DataLayout::NCHW, DataLayout::NHWC })); @@ -53,23 +63,19 @@ TEST_SUITE(Im2Col) using CLIm2Col = CLSynthetizeFunction; -// *INDENT-OFF* -// clang-format off DATA_TEST_CASE(Validate, framework::DatasetMode::ALL, zip(zip(zip( framework::dataset::make("InputInfo", { TensorInfo(TensorShape(10U, 12U, 2U), 1, DataType::U8), // Unsupported data type TensorInfo(TensorShape(10U, 12U, 2U), 1, DataType::F32), // Mismatching data type TensorInfo(TensorShape(10U, 12U, 2U), 1, DataType::QASYMM8), // Bias not supported with QASYMM8 - TensorInfo(TensorShape(10U, 12U, 2U), 1, DataType::QASYMM8), // Mismatching shapes TensorInfo(TensorShape(10U, 12U, 2U, 2U), 1, DataType::QASYMM8), }), framework::dataset::make("OutputInfo",{ TensorInfo(TensorShape(3U, 4U, 10U, 2U), 1, DataType::F16), TensorInfo(TensorShape(3U, 4U, 10U, 2U), 1, DataType::F16), TensorInfo(TensorShape(3U, 3U, 10U, 2U), 1, DataType::QASYMM8), - TensorInfo(TensorShape(3U, 4U, 10U, 2U), 1, DataType::QASYMM8), - TensorInfo(TensorShape(18U, 80U, 1U, 2U), 1, DataType::QASYMM8), + TensorInfo(TensorShape(18U, 80U, 2U, 1U), 1, DataType::QASYMM8), })), - framework::dataset::make("HasBias", { true, true, true, false, false })), - framework::dataset::make("Expected", { false, false, false, true, true })), + framework::dataset::make("HasBias", { true, true, true, false })), + framework::dataset::make("Expected", { false, false, false, true })), input_info, output_info, has_bias, expected) { @@ -83,16 +89,18 @@ template using CLIm2ColFixture = Im2ColValidationFixture; TEST_SUITE(Float) TEST_SUITE(FP32) -FIXTURE_DATA_TEST_CASE(RunSmall, CLIm2ColFixture, framework::DatasetMode::ALL, combine(combine(datasets::SmallShapes(), framework::dataset::make("DataType", DataType::F32)), - conv_args)) +FIXTURE_DATA_TEST_CASE(RunSmall, CLIm2ColFixture, framework::DatasetMode::ALL, combine(combine(combine(datasets::SmallShapes(), framework::dataset::make("DataType", DataType::F32)), + conv_args), + framework::dataset::make("ChannelsFirstOutputNHWC", true))) { // Validate output validate(CLAccessor(_target), _reference); } TEST_SUITE_END() -FIXTURE_DATA_TEST_CASE(RunLarge, CLIm2ColFixture, framework::DatasetMode::NIGHTLY, combine(combine(datasets::LargeShapes(), framework::dataset::make("DataType", DataType::F32)), - conv_args)) +FIXTURE_DATA_TEST_CASE(RunLarge, CLIm2ColFixture, framework::DatasetMode::NIGHTLY, combine(combine(combine(datasets::LargeShapes(), framework::dataset::make("DataType", DataType::F32)), + conv_args), + framework::dataset::make("ChannelsFirstOutputNHWC", true))) { // Validate output validate(CLAccessor(_target), _reference); @@ -101,14 +109,16 @@ FIXTURE_DATA_TEST_CASE(RunLarge, CLIm2ColFixture, framework::DatasetMode: #ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC TEST_SUITE(FP16) -FIXTURE_DATA_TEST_CASE(RunSmall, CLIm2ColFixture, framework::DatasetMode::ALL, combine(combine(datasets::SmallShapes(), framework::dataset::make("DataType", DataType::F16)), - conv_args)) +FIXTURE_DATA_TEST_CASE(RunSmall, CLIm2ColFixture, framework::DatasetMode::ALL, combine(combine(combine(datasets::SmallShapes(), framework::dataset::make("DataType", DataType::F16)), + conv_args), + framework::dataset::make("ChannelsFirstOutputNHWC", true))) { // Validate output validate(CLAccessor(_target), _reference); } -FIXTURE_DATA_TEST_CASE(RunLarge, CLIm2ColFixture, framework::DatasetMode::NIGHTLY, combine(combine(datasets::LargeShapes(), framework::dataset::make("DataType", DataType::F16)), - conv_args)) +FIXTURE_DATA_TEST_CASE(RunLarge, CLIm2ColFixture, framework::DatasetMode::NIGHTLY, combine(combine(combine(datasets::LargeShapes(), framework::dataset::make("DataType", DataType::F16)), + conv_args), + framework::dataset::make("ChannelsFirstOutputNHWC", true))) { // Validate output validate(CLAccessor(_target), _reference); @@ -120,14 +130,16 @@ TEST_SUITE_END() TEST_SUITE_END() TEST_SUITE(QASYMM8) -FIXTURE_DATA_TEST_CASE(RunSmall, CLIm2ColFixture, framework::DatasetMode::ALL, combine(combine(datasets::SmallShapes(), framework::dataset::make("DataType", DataType::QASYMM8)), - conv_args)) +FIXTURE_DATA_TEST_CASE(RunSmall, CLIm2ColFixture, framework::DatasetMode::ALL, combine(combine(combine(datasets::SmallShapes(), framework::dataset::make("DataType", DataType::QASYMM8)), + conv_args), + framework::dataset::make("ChannelsFirstOutputNHWC", true))) { // Validate output validate(CLAccessor(_target), _reference); } -FIXTURE_DATA_TEST_CASE(RunLarge, CLIm2ColFixture, framework::DatasetMode::NIGHTLY, combine(combine(datasets::LargeShapes(), framework::dataset::make("DataType", DataType::QASYMM8)), - conv_args)) +FIXTURE_DATA_TEST_CASE(RunLarge, CLIm2ColFixture, framework::DatasetMode::NIGHTLY, combine(combine(combine(datasets::LargeShapes(), framework::dataset::make("DataType", DataType::QASYMM8)), + conv_args), + framework::dataset::make("ChannelsFirstOutputNHWC", false))) { // Validate output validate(CLAccessor(_target), _reference); diff --git a/tests/validation/CL/LocallyConnected.cpp b/tests/validation/CL/LocallyConnected.cpp index 5381072131..dbfe4e269f 100644 --- a/tests/validation/CL/LocallyConnected.cpp +++ b/tests/validation/CL/LocallyConnected.cpp @@ -59,6 +59,7 @@ DATA_TEST_CASE(Validate, framework::DatasetMode::ALL, zip(zip(zip(zip(zip( TensorInfo(TensorShape(23U, 27U, 5U), 1, DataType::F32), // Mismatching shape input/bias TensorInfo(TensorShape(23U, 27U, 5U), 1, DataType::F32), // Mismatching shape input/output TensorInfo(TensorShape(23U, 27U, 5U), 1, DataType::F32), // Asymmetric padding + TensorInfo(TensorShape(23U, 27U, 5U), 1, DataType::F32), // Padding required TensorInfo(TensorShape(23U, 27U, 5U), 1, DataType::F32) }), framework::dataset::make("WeightsInfo",{ TensorInfo(TensorShape(3U, 3U, 5U, 21U, 275U), 1, DataType::F16), @@ -68,7 +69,8 @@ DATA_TEST_CASE(Validate, framework::DatasetMode::ALL, zip(zip(zip(zip(zip( TensorInfo(TensorShape(3U, 3U, 5U, 21U, 275U), 1, DataType::F32), TensorInfo(TensorShape(3U, 3U, 5U, 21U, 275U), 1, DataType::F32), TensorInfo(TensorShape(3U, 3U, 5U, 21U, 275U), 1, DataType::F32), - TensorInfo(TensorShape(3U, 3U, 5U, 21U, 275U), 1, DataType::F32) + TensorInfo(TensorShape(3U, 3U, 5U, 21U, 275U), 1, DataType::F32), + TensorInfo(TensorShape(1U, 3U, 5U, 21U, 575U), 1, DataType::F32) })), framework::dataset::make("BiasInfo", { TensorInfo(TensorShape(21U, 275U), 1, DataType::F32), TensorInfo(TensorShape(21U, 275U), 1, DataType::F16), @@ -77,7 +79,8 @@ DATA_TEST_CASE(Validate, framework::DatasetMode::ALL, zip(zip(zip(zip(zip( TensorInfo(TensorShape(21U, 274U), 1, DataType::F32), TensorInfo(TensorShape(21U, 275U), 1, DataType::F32), TensorInfo(TensorShape(21U, 275U), 1, DataType::F32), - TensorInfo(TensorShape(21U, 275U), 1, DataType::F32) + TensorInfo(TensorShape(21U, 275U), 1, DataType::F32), + TensorInfo(TensorShape(21U, 575U), 1, DataType::F32) })), framework::dataset::make("OutputInfo", { TensorInfo(TensorShape(11U, 25U, 21U), 1, DataType::F32), TensorInfo(TensorShape(11U, 25U, 21U), 1, DataType::F32), @@ -86,7 +89,8 @@ DATA_TEST_CASE(Validate, framework::DatasetMode::ALL, zip(zip(zip(zip(zip( TensorInfo(TensorShape(11U, 25U, 21U), 1, DataType::F32), TensorInfo(TensorShape(11U, 25U, 22U), 1, DataType::F32), TensorInfo(TensorShape(11U, 25U, 21U), 1, DataType::F32), - TensorInfo(TensorShape(11U, 25U, 21U), 1, DataType::F32) + TensorInfo(TensorShape(11U, 25U, 21U), 1, DataType::F32), + TensorInfo(TensorShape(23U, 25U, 21U), 1, DataType::F32) })), framework::dataset::make("PadStride", { PadStrideInfo(2, 1, 0, 0), PadStrideInfo(2, 1, 0, 0), @@ -94,10 +98,11 @@ DATA_TEST_CASE(Validate, framework::DatasetMode::ALL, zip(zip(zip(zip(zip( PadStrideInfo(2, 1, 0, 0), PadStrideInfo(2, 1, 0, 0), PadStrideInfo(2, 1, 0, 0), - PadStrideInfo(2, 1, 1, 0, 0, 0, DimensionRoundingType::FLOOR), - PadStrideInfo(2, 1, 0, 0) + PadStrideInfo(2, 1, 1, 0), + PadStrideInfo(2, 1, 0, 0), + PadStrideInfo(1, 1, 0, 0) })), - framework::dataset::make("Expected", { false, false, false, false, false, false, false, true })), + framework::dataset::make("Expected", { false, false, false, false, false, false, false, false, true })), input_info, weights_info, bias_info, output_info, conv_info, expected) { bool is_valid = bool(CLLocallyConnectedLayer::validate(&input_info.clone()->set_is_resizable(false), -- cgit v1.2.1