/* * Copyright (c) 2022-2024 Arm Limited. * * SPDX-License-Identifier: MIT * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to * deal in the Software without restriction, including without limitation the * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or * sell copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in all * copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include "arm_compute/dynamic_fusion/sketch/gpu/operators/GpuResize.h" #include "tests/CL/CLAccessor.h" #include "tests/datasets/ScaleValidationDataset.h" #include "tests/framework/Asserts.h" #include "tests/framework/Fixture.h" #include "tests/framework/Macros.h" #include "tests/validation/fixtures/dynamic_fusion/operators/ResizeFixture.h" #include "tests/validation/Validation.h" using namespace arm_compute::experimental::dynamic_fusion; namespace arm_compute { namespace test { namespace validation { namespace { using datasets::ScaleAlignCornersSamplingPolicySet; using datasets::ScaleInterpolationPolicySet; using datasets::ScaleSamplingPolicySet; using datasets::ScaleShapesBaseDataSet; /** We consider vector size in byte 16 since the maximum size of * a vector used by @ref CLScaleKernel is currently 16-byte (float4). */ constexpr uint32_t vector_byte = 16; template constexpr uint32_t num_elements_per_vector() { return vector_byte / sizeof(T); } /** Quantization information data set */ const auto QuantizationInfoSet = framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.5f, -1), }); /** Tolerance */ constexpr float tolerance_f32_absolute(0.001f); RelativeTolerance tolerance_f32(0.05); constexpr float abs_tolerance_f16(0.1f); RelativeTolerance tolerance_f16(half(0.1)); constexpr float tolerance_num_f32(0.01f); } // namespace TEST_SUITE(CL) TEST_SUITE(DYNAMIC_FUSION) TEST_SUITE(RESIZE) TEST_SUITE(Validate) const auto default_input_shape = TensorShape{2, 3, 3, 2}; const auto default_output_shape = TensorShape{4, 6, 3, 2}; constexpr auto default_data_type = DataType::U8; constexpr auto default_data_layout = DataLayout::NHWC; TEST_CASE(NullPtr, framework::DatasetMode::ALL) { const TensorInfo input_info = TensorInfo{default_input_shape, 1, default_data_type, default_data_layout}; const TensorInfo output_info = TensorInfo{default_output_shape, 1, default_data_type, default_data_layout}; CLCompileContext cl_compile_ctx = CLKernelLibrary::get().get_compile_context(); GpuWorkloadContext context = GpuWorkloadContext{&cl_compile_ctx}; GpuWorkloadSketch sketch{&context}; // nullptr is given as input Status status = GpuResize::validate_op(sketch, nullptr, ResizeAttributes()); ARM_COMPUTE_EXPECT(bool(status) == false, framework::LogLevel::ERRORS); } TEST_CASE(SupportDataType, framework::DatasetMode::ALL) { const std::map supported_data_types = { { DataType::U8, false }, { DataType::S8, false }, { DataType::QSYMM8, false }, { DataType::QASYMM8, false }, { DataType::QASYMM8_SIGNED, false }, { DataType::QSYMM8_PER_CHANNEL, false }, { DataType::U16, false }, { DataType::S16, false }, { DataType::QSYMM16, false }, { DataType::QASYMM16, false }, { DataType::U32, false }, { DataType::S32, false }, { DataType::U64, false }, { DataType::S64, false }, { DataType::BFLOAT16, false }, { DataType::F16, true }, { DataType::F32, true }, { DataType::F64, false }, { DataType::SIZET, false }, }; for (auto &kv : supported_data_types) { const TensorInfo input_info = TensorInfo{default_input_shape, 1, kv.first, default_data_layout}; CLCompileContext cl_compile_ctx = CLKernelLibrary::get().get_compile_context(); GpuWorkloadContext context = GpuWorkloadContext{&cl_compile_ctx}; GpuWorkloadSketch sketch{&context}; const ITensorInfo *sketch_input_info = context.create_tensor_info(input_info); ResizeAttributes attributes; attributes.output_width(default_output_shape[0]); // shape is not important unless it's empty attributes.output_height(default_output_shape[1]); Status status = GpuResize::validate_op(sketch, sketch_input_info, attributes); ARM_COMPUTE_EXPECT(bool(status) == kv.second, framework::LogLevel::ERRORS); } } TEST_CASE(MismatchingDataType, framework::DatasetMode::ALL) { constexpr DataType non_default_data_type = DataType::F32; const TensorInfo input_info = TensorInfo{default_input_shape, 1, default_data_type, default_data_layout}; const TensorInfo output_info = TensorInfo{default_output_shape, 1, non_default_data_type, default_data_layout}; CLCompileContext cl_compile_ctx = CLKernelLibrary::get().get_compile_context(); GpuWorkloadContext context = GpuWorkloadContext{&cl_compile_ctx}; GpuWorkloadSketch sketch{&context}; const ITensorInfo *sketch_input_info = context.create_tensor_info(input_info); Status status = GpuResize::validate_op(sketch, sketch_input_info, ResizeAttributes()); ARM_COMPUTE_EXPECT(bool(status) == false, framework::LogLevel::ERRORS); } TEST_CASE(AlignedCornerNotSupported, framework::DatasetMode::ALL) { // Aligned corners require sampling policy to be TOP_LEFT. constexpr InterpolationPolicy interpolation_policy = InterpolationPolicy::BILINEAR; constexpr bool align_corners = true; constexpr SamplingPolicy sampling_policy = SamplingPolicy::CENTER; const TensorInfo input_info = TensorInfo{default_input_shape, 1, default_data_type, default_data_layout}; const TensorInfo output_info = TensorInfo{default_output_shape, 1, default_data_type, default_data_layout}; CLCompileContext cl_compile_ctx = CLKernelLibrary::get().get_compile_context(); GpuWorkloadContext context = GpuWorkloadContext{&cl_compile_ctx}; GpuWorkloadSketch sketch{&context}; const ITensorInfo *sketch_input_info = context.create_tensor_info(input_info); ResizeAttributes attributes{}; attributes.interpolation_policy(interpolation_policy).sampling_policy(sampling_policy).align_corners(align_corners); Status status = GpuResize::validate_op(sketch, sketch_input_info, attributes); ARM_COMPUTE_EXPECT(bool(status) == false, framework::LogLevel::ERRORS); } TEST_CASE(UnsupportedInterpolationPolicy, framework::DatasetMode::ALL) { const TensorInfo input_info = TensorInfo{TensorShape(28U, 33U, 2U), 1, DataType::F32, default_data_layout}; const TensorInfo output_info = TensorInfo{TensorShape(26U, 21U, 2U), 1, DataType::F32, default_data_layout}; constexpr auto interpolation_policy = InterpolationPolicy::AREA; CLCompileContext cl_compile_ctx = CLKernelLibrary::get().get_compile_context(); GpuWorkloadContext context = GpuWorkloadContext{&cl_compile_ctx}; GpuWorkloadSketch sketch{&context}; const ITensorInfo *sketch_input_info = context.create_tensor_info(input_info); ResizeAttributes attributes{}; attributes.interpolation_policy(interpolation_policy); Status status = GpuResize::validate_op(sketch, sketch_input_info, attributes); ARM_COMPUTE_EXPECT(bool(status) == false, framework::LogLevel::ERRORS); } TEST_CASE(UnsupportedLayout, framework::DatasetMode::ALL) { const TensorInfo input_info = TensorInfo{default_input_shape, 1, default_data_type, DataLayout::NCHW}; const TensorInfo output_info = TensorInfo{default_output_shape, 1, default_data_type, DataLayout::NCHW}; constexpr auto interpolation_policy = InterpolationPolicy::BILINEAR; CLCompileContext cl_compile_ctx = CLKernelLibrary::get().get_compile_context(); GpuWorkloadContext context = GpuWorkloadContext{&cl_compile_ctx}; GpuWorkloadSketch sketch{&context}; const ITensorInfo *sketch_input_info = context.create_tensor_info(input_info); ResizeAttributes attributes{}; attributes.interpolation_policy(interpolation_policy); Status status = GpuResize::validate_op(sketch, sketch_input_info, attributes); ARM_COMPUTE_EXPECT(bool(status) == false, framework::LogLevel::ERRORS); } TEST_SUITE_END() // Validate template using DynamicFusionResizeFixture = DynamicFusionResizeValidationFixture; TEST_SUITE(Float) TEST_SUITE(FP32) const auto f32_shape = combine((SCALE_PRECOMMIT_SHAPE_DATASET(num_elements_per_vector())), framework::dataset::make("DataType", DataType::F32)); FIXTURE_DATA_TEST_CASE(Run, DynamicFusionResizeFixture, framework::DatasetMode::ALL, ASSEMBLE_DATASET_DYNAMIC_FUSION(f32_shape, ScaleSamplingPolicySet)) { //Create valid region TensorInfo src_info(_shape, 1, _data_type); const ValidRegion valid_region = calculate_valid_region_scale(src_info, _reference.shape(), _interpolation_policy, _sampling_policy, false); // Validate output validate(CLAccessor(_target), _reference, valid_region, tolerance_f32, tolerance_num_f32, tolerance_f32_absolute); } FIXTURE_DATA_TEST_CASE(RunAlignCorners, DynamicFusionResizeFixture, framework::DatasetMode::ALL, ASSEMBLE_DATASET_DYNAMIC_FUSION(f32_shape, ScaleAlignCornersSamplingPolicySet)) { //Create valid region TensorInfo src_info(_shape, 1, _data_type); const ValidRegion valid_region = calculate_valid_region_scale(src_info, _reference.shape(), _interpolation_policy, _sampling_policy, false); // Validate output validate(CLAccessor(_target), _reference, valid_region, tolerance_f32, tolerance_num_f32, tolerance_f32_absolute); } const auto f32_nightly_shape = combine((SCALE_NIGHTLY_SHAPE_DATASET(num_elements_per_vector())), framework::dataset::make("DataType", DataType::F32)); FIXTURE_DATA_TEST_CASE(RunNightly, DynamicFusionResizeFixture, framework::DatasetMode::NIGHTLY, ASSEMBLE_DATASET_DYNAMIC_FUSION(f32_nightly_shape, ScaleSamplingPolicySet)) { //Create valid region TensorInfo src_info(_shape, 1, _data_type); const ValidRegion valid_region = calculate_valid_region_scale(src_info, _reference.shape(), _interpolation_policy, _sampling_policy, false); // Validate output validate(CLAccessor(_target), _reference, valid_region, tolerance_f32, tolerance_num_f32, tolerance_f32_absolute); } FIXTURE_DATA_TEST_CASE(RunNightlyAlignCorners, DynamicFusionResizeFixture, framework::DatasetMode::NIGHTLY, ASSEMBLE_DATASET_DYNAMIC_FUSION(f32_nightly_shape, ScaleAlignCornersSamplingPolicySet)) { //Create valid region TensorInfo src_info(_shape, 1, _data_type); const ValidRegion valid_region = calculate_valid_region_scale(src_info, _reference.shape(), _interpolation_policy, _sampling_policy, false); // Validate output validate(CLAccessor(_target), _reference, valid_region, tolerance_f32, tolerance_num_f32, tolerance_f32_absolute); } TEST_SUITE_END() // FP32 TEST_SUITE(FP16) const auto f16_shape = combine((SCALE_PRECOMMIT_SHAPE_DATASET(num_elements_per_vector())), framework::dataset::make("DataType", DataType::F16)); FIXTURE_DATA_TEST_CASE(Run, DynamicFusionResizeFixture, framework::DatasetMode::ALL, ASSEMBLE_DATASET_DYNAMIC_FUSION(f16_shape, ScaleSamplingPolicySet)) { //Create valid region TensorInfo src_info(_shape, 1, _data_type); const ValidRegion valid_region = calculate_valid_region_scale(src_info, _reference.shape(), _interpolation_policy, _sampling_policy, false); // Validate output validate(CLAccessor(_target), _reference, valid_region, tolerance_f16, 0.0f, abs_tolerance_f16); } FIXTURE_DATA_TEST_CASE(RunAlignCorners, DynamicFusionResizeFixture, framework::DatasetMode::ALL, ASSEMBLE_DATASET_DYNAMIC_FUSION(f16_shape, ScaleAlignCornersSamplingPolicySet)) { //Create valid region TensorInfo src_info(_shape, 1, _data_type); const ValidRegion valid_region = calculate_valid_region_scale(src_info, _reference.shape(), _interpolation_policy, _sampling_policy, false); // Validate output validate(CLAccessor(_target), _reference, valid_region, tolerance_f16, 0.0f, abs_tolerance_f16); } const auto f16_nightly_shape = combine((SCALE_NIGHTLY_SHAPE_DATASET(num_elements_per_vector())), framework::dataset::make("DataType", DataType::F16)); FIXTURE_DATA_TEST_CASE(RunNightly, DynamicFusionResizeFixture, framework::DatasetMode::NIGHTLY, ASSEMBLE_DATASET_DYNAMIC_FUSION(f16_nightly_shape, ScaleSamplingPolicySet)) { //Create valid region TensorInfo src_info(_shape, 1, _data_type); const ValidRegion valid_region = calculate_valid_region_scale(src_info, _reference.shape(), _interpolation_policy, _sampling_policy, false); // Validate output validate(CLAccessor(_target), _reference, valid_region, tolerance_f16, 0.0f, abs_tolerance_f16); } FIXTURE_DATA_TEST_CASE(RunNightlyAlignCorners, DynamicFusionResizeFixture, framework::DatasetMode::NIGHTLY, ASSEMBLE_DATASET_DYNAMIC_FUSION(f16_nightly_shape, ScaleAlignCornersSamplingPolicySet)) { //Create valid region TensorInfo src_info(_shape, 1, _data_type); const ValidRegion valid_region = calculate_valid_region_scale(src_info, _reference.shape(), _interpolation_policy, _sampling_policy, false); // Validate output validate(CLAccessor(_target), _reference, valid_region, tolerance_f16, 0.0f, abs_tolerance_f16); } TEST_SUITE_END() // FP16 TEST_SUITE_END() // Float TEST_SUITE_END() // RESIZE TEST_SUITE_END() // DYNAMIC_FUSION TEST_SUITE_END() // CL } // namespace validation } // namespace test } // namespace arm_compute