From 282f324f4d6aeff172d447be6dfdf54c3d38b235 Mon Sep 17 00:00:00 2001 From: SiCongLi Date: Tue, 24 Nov 2020 15:24:16 +0000 Subject: COMPMID-3982 Add QASYMM8 support to gemm tuner (reshaped and reshaped_only_rhs) Signed-off-by: SiCongLi Change-Id: Ic6492cfd2701374d837ac53d51b0ddc07e6b1fd1 Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/4550 Tested-by: Arm Jenkins Reviewed-by: Gian Marco Iodice Comments-Addressed: Arm Jenkins --- ...aped_rhs_only_fused_output_stage_fixedpoint.cpp | 364 +++++++++++++++++++++ 1 file changed, 364 insertions(+) create mode 100644 examples/gemm_tuner/cl_gemmlowp_reshaped_rhs_only_fused_output_stage_fixedpoint.cpp (limited to 'examples/gemm_tuner/cl_gemmlowp_reshaped_rhs_only_fused_output_stage_fixedpoint.cpp') diff --git a/examples/gemm_tuner/cl_gemmlowp_reshaped_rhs_only_fused_output_stage_fixedpoint.cpp b/examples/gemm_tuner/cl_gemmlowp_reshaped_rhs_only_fused_output_stage_fixedpoint.cpp new file mode 100644 index 0000000000..c6818e48b0 --- /dev/null +++ b/examples/gemm_tuner/cl_gemmlowp_reshaped_rhs_only_fused_output_stage_fixedpoint.cpp @@ -0,0 +1,364 @@ +/* + * Copyright (c) 2020 Arm Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ +#ifndef ARM_COMPUTE_CL /* Needed by Utils.cpp to handle OpenCL exceptions properly */ +#error "This example needs to be built with -DARM_COMPUTE_CL" +#endif /* ARM_COMPUTE_CL */ + +#include "CommonGemmExampleOptions.h" +#include "GemmTunerHelpers.h" +#include "arm_compute/core/Helpers.h" +#include "arm_compute/core/KernelDescriptors.h" +#include "arm_compute/core/KernelDescriptors.h" +#include "arm_compute/core/Types.h" +#include "arm_compute/core/utils/misc/ShapeCalculator.h" +#include "arm_compute/core/utils/quantization/AsymmHelpers.h" +#include "arm_compute/runtime/CL/CLScheduler.h" +#include "arm_compute/runtime/CL/CLTuner.h" +#include "src/core/CL/kernels/CLGEMMLowpMatrixMultiplyReshapedOnlyRHSKernel.h" +#include "src/core/CL/kernels/CLGEMMLowpReductionKernel.h" +#include "tests/CL/Helper.h" +#include "utils/Utils.h" +#include "utils/command_line/CommandLineOptions.h" +#include "utils/command_line/CommandLineParser.h" + +#include +#include + +using namespace arm_compute; +using namespace utils; +using namespace arm_compute::misc::shape_calculator; +using namespace gemm_tuner; + +namespace +{ +/** Structure holding all tunable gemm configs specific to this example/strategy */ +struct GemmConfigs +{ + size_t m0{ 4 }; /**< Number of rows processed by the matrix multiplication */ + size_t n0{ 4 }; /**< Number of columns processed by the matrix multiplication */ + size_t k0{ 4 }; /**< Number of partial accumulations performed by the matrix multiplication */ + size_t h0{ 1 }; /**< Number of horizontal blocks of size (k0xn0) stored on the same output row */ + bool interleave_rhs{ true }; /**< Interleave rhs matrix */ + bool transpose_rhs{ true }; /**< Transpose rhs matrix */ +}; + +/** Formatted output of the GemmConfigs type + * + * @param[out] os Output stream. + * @param[in] configs Tunable configurations to output + * + * @return Modified output stream. + */ +::std::ostream &operator<<(::std::ostream &os, const GemmConfigs &configs) +{ + std::string false_str = std::string("false"); + std::string true_str = std::string("true"); + + os << "m0 : " << configs.m0 << std::endl; + os << "n0 : " << configs.n0 << std::endl; + os << "k0 : " << configs.k0 << std::endl; + os << "h0 : " << configs.h0 << std::endl; + os << "interleave_rhs : " << (configs.interleave_rhs ? true_str : false_str) << std::endl; + os << "transpose_rhs : " << (configs.transpose_rhs ? true_str : false_str) << std::endl; + return os; +} + +/** Command line options for gemm configs */ +class GemmConfigOptions +{ +public: + /** Constructor + * + * @param[in,out] parser A parser on which "parse()" hasn't been called yet. + */ + GemmConfigOptions(CommandLineParser &parser) + : m0(parser.add_positional_option>("m0", 4)), + n0(parser.add_positional_option>("n0", 4)), + k0(parser.add_positional_option>("k0", 4)), + h0(parser.add_positional_option>("h0", 1)), + interleave_rhs(parser.add_positional_option>("interleave_rhs", 1)), + transpose_rhs(parser.add_positional_option>("transpose_rhs", 1)) + { + m0->set_help("Number of rows processed by the matrix multiplication"); + n0->set_help("Number of columns processed by the matrix multiplication"); + k0->set_help("Number of partial accumulations performed by the matrix multiplication"); + h0->set_help("Number of horizontal blocks of size (k0xn0) stored on the same output row"); + interleave_rhs->set_help("Interleave rhs matrix (1) / Do not interleave rhs matrix (0)"); + transpose_rhs->set_help("Transpose rhs matrix (1) / Do not transpose rhs matrix (0)"); + } + /** Prevent instances of this class from being copied (As this class contains pointers) */ + GemmConfigOptions(const GemmConfigOptions &) = delete; + /** Prevent instances of this class from being copied (As this class contains pointers) */ + GemmConfigOptions &operator=(const GemmConfigOptions &) = delete; + /** Allow instances of this class to be moved */ + GemmConfigOptions(GemmConfigOptions &&) = default; + /** Allow instances of this class to be moved */ + GemmConfigOptions &operator=(GemmConfigOptions &&) = default; + /** Default destructor */ + ~GemmConfigOptions() = default; + + SimpleOption *m0; /**< Number of rows processed by the matrix multiplication option */ + SimpleOption *n0; /**< Number of columns processed by the matrix multiplication option */ + SimpleOption *k0; /**< Number of partial accumulations performed by the matrix multiplication option */ + SimpleOption *h0; /**< Number of horizontal blocks of size (k0xn0) stored on the same output row option */ + SimpleOption *interleave_rhs; /**< Interleave rhs matrix option (1 enable; 0 disable) */ + SimpleOption *transpose_rhs; /**< Transpose rhs matrix option (1 enable; 0 disable) */ +}; + +/** Consumes the gemm configuration options and creates a structure containing all information + * + * @param[in] options Options to consume + * + * @return Structure containing the gemm configurations + */ +GemmConfigs consume_gemm_configs(const GemmConfigOptions &options) +{ + GemmConfigs configs; + configs.m0 = options.m0->value(); + configs.n0 = options.n0->value(); + configs.k0 = options.k0->value(); + configs.h0 = options.h0->value(); + configs.interleave_rhs = options.interleave_rhs->value() != 0; + configs.transpose_rhs = options.transpose_rhs->value() != 0; + return configs; +} + +} // namespace + +using CLGEMMLowpMatrixMultiplyReshapedOnlyRHS = test::CLSynthetizeFunction; +using CLGEMMLowpMatrixAReduction = test::CLSynthetizeFunction; + +class CLGEMMLowpMatrixMultiplyReshapedOnlyRHSFusedOutputStageFixedpointExample : public Example +{ +public: + bool do_setup(int argc, char **argv) override + { + // Default parameters + CommonGemmExampleParams params; + GemmConfigs configs; + + // Parse command line options + CommandLineParser parser; + CommonGemmExampleOptions param_options(parser); + GemmConfigOptions config_options(parser); + + parser.parse(argc, argv); + if(param_options.help->is_set() && param_options.help->value()) + { + parser.print_help(argv[0]); + return false; + } + if(!parser.validate()) + { + // Invalid arguments. Use default parameters and configs + std::cerr << "Invalid arguments." << std::endl; + parser.print_help(argv[0]); + std::cerr << "Falling back to default parameters and configs" << std::endl; + } + else + { + params = consume_common_gemm_example_parameters(param_options); + configs = consume_gemm_configs(config_options); + } + + std::cout << "Gemm parameters:" << std::endl; + std::cout << params << std::endl; + std::cout << "Gemm configurations:" << std::endl; + std::cout << configs << std::endl; + + CLScheduler::get().default_init(&tuner); + + lhs.allocator()->init(TensorInfo(TensorShape(params.K, params.M, params.B), 1, params.data_type)); + rhs.allocator()->init(TensorInfo(TensorShape(params.N, params.K, params.B), 1, params.data_type)); + bias.allocator()->init(TensorInfo(TensorShape(params.N, 1, params.B), 1, DataType::S32)); + dst.allocator()->init(TensorInfo(TensorShape(params.N, params.M, params.B), 1, params.data_type)); + + // Set arbitrary quantization information (non-zero offset to ensure offset contribution stage is included) + // Could be extended in the future to include a user-controlled option for offset == 0 + lhs.info()->set_quantization_info({ 0.012, 3 }); + rhs.info()->set_quantization_info({ 0.012, 3 }); + bias.info()->set_quantization_info({ 0.012, 3 }); + dst.info()->set_quantization_info({ 0.012, 3 }); + + GEMMLHSMatrixInfo lhs_info; + lhs_info.m0 = configs.m0; + lhs_info.k0 = configs.k0; + + GEMMRHSMatrixInfo rhs_info; + rhs_info.n0 = configs.n0; + rhs_info.k0 = configs.k0; + rhs_info.h0 = configs.h0; + rhs_info.interleave = configs.interleave_rhs; + rhs_info.transpose = configs.transpose_rhs; + rhs_info.export_to_cl_image = false; // CL image not supported for quantized cases yet + + rhs_reshaped.allocator()->init(TensorInfo(compute_rhs_reshaped_shape(*rhs.info(), rhs_info), 1, params.data_type)); + if(rhs_info.export_to_cl_image) + { + examples::gemm_tuner_helpers::update_padding_for_cl_image(rhs_reshaped.info()); + } + + // Configure output stage for quantized case + GEMMLowpOutputStageInfo gemmlowp_output_stage; + gemmlowp_output_stage.type = GEMMLowpOutputStageType::QUANTIZE_DOWN_FIXEDPOINT; + gemmlowp_output_stage.output_data_type = dst.info()->data_type(); + gemmlowp_output_stage.gemmlowp_offset = 0; + { + const int idx_kernels = get_data_layout_dimension_index(lhs.info()->data_layout(), DataLayoutDimension::BATCHES); + gemmlowp_output_stage.is_quantized_per_channel = false; + // Num_filters is 1 unless quantized type is of per_channel type. Could be extended in the future to support per-channel quantization. + const unsigned int num_filters = 1; + + dst_multipliers.allocator()->init(TensorInfo(TensorShape(num_filters), 1, DataType::S32)); + dst_shifts.allocator()->init(TensorInfo(TensorShape(num_filters), 1, DataType::S32)); + + gemmlowp_output_stage.gemmlowp_multipliers.resize(num_filters); + gemmlowp_output_stage.gemmlowp_shifts.resize(num_filters); + quantization::compute_quantized_multipliers_and_shifts(lhs.info(), + rhs.info(), + dst.info(), + idx_kernels, + gemmlowp_output_stage.gemmlowp_multipliers.data(), + gemmlowp_output_stage.gemmlowp_shifts.data()); + gemmlowp_output_stage.gemmlowp_multiplier = gemmlowp_output_stage.gemmlowp_multipliers[0]; + gemmlowp_output_stage.gemmlowp_shift = gemmlowp_output_stage.gemmlowp_shifts[0]; + + // No fused activation + PixelValue min_val{}; + PixelValue max_val{}; + std::tie(min_val, max_val) = get_min_max(dst.info()->data_type()); + + auto min_activation = min_val.get(); + auto max_activation = max_val.get(); + + // Set the GEMMLowp output stage info + gemmlowp_output_stage.gemmlowp_offset = dst.info()->quantization_info().uniform().offset; + gemmlowp_output_stage.gemmlowp_min_bound = min_activation; + gemmlowp_output_stage.gemmlowp_max_bound = max_activation; + } + + GEMMKernelInfo gemm_info; + gemm_info.m = params.M; + gemm_info.n = params.N; + gemm_info.k = params.K; + gemm_info.depth_output_gemm3d = 0; + gemm_info.reinterpret_input_as_3d = false; + gemm_info.broadcast_bias = true; + gemm_info.fp_mixed_precision = false; + gemm_info.has_pad_y = false; + gemm_info.mult_transpose1xW_width = configs.h0; + gemm_info.lhs_info = lhs_info; + gemm_info.rhs_info = rhs_info; + gemm_info.a_offset = lhs.info()->quantization_info().uniform().offset; + gemm_info.b_offset = rhs.info()->quantization_info().uniform().offset; + gemm_info.output_stage = gemmlowp_output_stage; + + // Initialize Matrix A reduction kernel only if _b_offset is not equal to 0 + if(gemm_info.b_offset != 0) + { + const TensorInfo info_vector_sum_row(compute_reductionB_shape(*lhs.info()), 1, DataType::S32); + vector_sum_row.allocator()->init(info_vector_sum_row); + + mtx_a_reduction = support::cpp14::make_unique(); + + if(!mtx_a_reduction->validate(lhs.info(), vector_sum_row.info(), GEMMLowpReductionKernelInfo{})) + { + std::cerr << "Invalid arguments for CLGEMMLowpMatrixAReductionKernel." << std::endl; + return false; + } + + mtx_a_reduction->configure(&lhs, &vector_sum_row, GEMMLowpReductionKernelInfo{}); + } + // Initialize matrix B reduction kernel only if _a_offset is not equal to 0 + if(gemm_info.a_offset != 0) + { + const TensorInfo info_vector_sum_col(compute_reductionA_shape(*rhs.info()), 1, DataType::S32); + vector_sum_col.allocator()->init(info_vector_sum_col); + // There's no need for a Matrix B reduction kernel as this is assumed to be run only once in the prepare stage + } + + // Validate argments + if(!gemm.validate(lhs.info(), rhs_reshaped.info(), dst.info(), gemm_info, gemm_info.a_offset == 0 ? nullptr : vector_sum_col.info(), + gemm_info.b_offset == 0 ? nullptr : vector_sum_row.info(), bias.info(), dst_multipliers.info(), dst_shifts.info())) + { + std::cerr << "Invalid arguments for CLGEMMLowpMatrixMultiplyReshapedOnlyRHSKernel." << std::endl; + return false; + } + + // Configure function + gemm.configure(&lhs, &rhs_reshaped, &dst, gemm_info, gemm_info.a_offset == 0 ? nullptr : &vector_sum_col, gemm_info.b_offset == 0 ? nullptr : &vector_sum_row, &bias, &dst_multipliers, &dst_shifts); + + // Allocate tensors + lhs.allocator()->allocate(); + rhs.allocator()->allocate(); + rhs_reshaped.allocator()->allocate(); + bias.allocator()->allocate(); + dst.allocator()->allocate(); + vector_sum_col.allocator()->allocate(); + vector_sum_row.allocator()->allocate(); + dst_multipliers.allocator()->allocate(); + dst_shifts.allocator()->allocate(); + + return true; + } + void do_run() override + { + if(mtx_a_reduction != nullptr) + { + mtx_a_reduction->run(); + } + gemm.run(); + + // Make sure all the OpenCL jobs are done executing: + CLScheduler::get().sync(); + } + + void do_teardown() override + { + } + +private: + CLTensor lhs{}; + CLTensor rhs{}; + CLTensor rhs_reshaped{}; + CLTensor bias{}; + CLTensor dst{}; + CLTensor vector_sum_col{}; + CLTensor vector_sum_row{}; + CLTensor dst_multipliers{}; + CLTensor dst_shifts{}; + CLTuner tuner{}; + CLGEMMLowpMatrixMultiplyReshapedOnlyRHS gemm{}; + std::unique_ptr mtx_a_reduction{ nullptr }; +}; + +/** Main test program for gemmlowp reshaped rhs only with fused output stage fixedpoint + * + * @param[in] argc Number of arguments + * @param[in] argv Arguments ( [optional] M, [optional] N, [optional] K, [optional] B, [optional] m0, [optional] n0, [optional] k0, [optional] h0, [optional] interleave_rhs, [optional] transpose_rhs ) + */ +int main(int argc, char **argv) +{ + return run_example(argc, argv); +} -- cgit v1.2.1