From a5577dbe4735656e488aa7a47261b993e49b6562 Mon Sep 17 00:00:00 2001 From: Viet-Hoa Do Date: Thu, 13 Jul 2023 16:38:28 +0100 Subject: Fix dynamic fusion compilation error Resolves: COMPMID-6393 Signed-off-by: Viet-Hoa Do Change-Id: Idc0880a964f2827bf5bf267b72fe7db9ce116f15 Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/9919 Reviewed-by: Pablo Marquez Tello Comments-Addressed: Arm Jenkins Benchmark: Arm Jenkins Tested-by: Arm Jenkins --- .../fixtures/dynamic_fusion/gpu/cl/DepthwiseConv2dFixture.h | 2 -- tests/validation/fixtures/dynamic_fusion/gpu/cl/DirectConv2dFixture.h | 4 ---- .../fixtures/dynamic_fusion/gpu/cl/ElementwiseBinaryFixture.h | 4 ---- tests/validation/fixtures/dynamic_fusion/gpu/cl/Pool2dFixture.h | 4 ---- .../validation/fixtures/dynamic_fusion/operators/ActivationFixture.h | 3 --- tests/validation/fixtures/dynamic_fusion/operators/CastFixture.h | 1 - tests/validation/fixtures/dynamic_fusion/operators/ClampFixture.h | 1 - tests/validation/fixtures/dynamic_fusion/operators/MulFixture.h | 4 ---- tests/validation/fixtures/dynamic_fusion/operators/ReshapeFixture.h | 1 - tests/validation/fixtures/dynamic_fusion/operators/ResizeFixture.h | 3 --- tests/validation/fixtures/dynamic_fusion/operators/SoftmaxFixture.h | 2 -- 11 files changed, 29 deletions(-) (limited to 'tests/validation/fixtures') diff --git a/tests/validation/fixtures/dynamic_fusion/gpu/cl/DepthwiseConv2dFixture.h b/tests/validation/fixtures/dynamic_fusion/gpu/cl/DepthwiseConv2dFixture.h index bea1d9bf4b..6498a06e03 100644 --- a/tests/validation/fixtures/dynamic_fusion/gpu/cl/DepthwiseConv2dFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/gpu/cl/DepthwiseConv2dFixture.h @@ -60,7 +60,6 @@ public: || std::is_same::type, int8_t>::value, int32_t, T >::type; // If T: uint8_t or int8_t then TBias: int32_t, otherwise TBias: T - template void setup(TensorShape input_shape, Size2D kernel_size, const PadStrideInfo &pad_stride, const Size2D &dilation, const unsigned int depth_multiplier, const DataType data_type, const DataLayout data_layout) { @@ -213,7 +212,6 @@ template { public: - template void setup(TensorShape input_shape, Size2D kernel_size, const PadStrideInfo &info, const Size2D &dilation, const unsigned int depth_multiplier, DataType data_type, DataLayout data_layout) { DynamicFusionGpuDepthwiseConv2dValidationGenericFixture::setup(input_shape, kernel_size, info, dilation, diff --git a/tests/validation/fixtures/dynamic_fusion/gpu/cl/DirectConv2dFixture.h b/tests/validation/fixtures/dynamic_fusion/gpu/cl/DirectConv2dFixture.h index 81dfc2b8e2..e30a564930 100644 --- a/tests/validation/fixtures/dynamic_fusion/gpu/cl/DirectConv2dFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/gpu/cl/DirectConv2dFixture.h @@ -88,7 +88,6 @@ public: || std::is_same::type, int8_t>::value, int32_t, T >::type; // If T: uint8_t or int8_t then TBias: int32_t, otherwise TBias: T - template void setup(TensorShape input_shape, TensorShape weights_shape, TensorShape bias_shape, TensorShape output_shape, const PadStrideInfo &info, const Size2D &dilation, DataType data_type, DataLayout data_layout, QuantizationInfo quantization_info, QuantizationInfo weight_quantization_info) { @@ -203,7 +202,6 @@ template { public: - template void setup(TensorShape input_shape, TensorShape weights_shape, TensorShape output_shape, TensorShape bias_shape, const PadStrideInfo &info, const Size2D &dialation, DataType data_type, DataLayout data_layout, QuantizationInfo quantization_info) { @@ -222,7 +220,6 @@ class DynamicFusionDirectConv2dValidationGenericFixture : public framework::Fixt public: using TBias = typename std::conditional < std::is_same::value || std::is_same::value, int32_t, T >::type; - template void setup(TensorShape input_shape, int stride_x, int stride_y, int pad_x, int pad_y, unsigned int kernel_size, unsigned int num_kernels, DataType data_type, QuantizationInfo quantization_info, DataLayout data_layout) { @@ -341,7 +338,6 @@ template { public: - template void setup(TensorShape input_shape, int stride_x, int stride_y, int pad_x, int pad_y, unsigned int kernel_size, unsigned int num_kernels, DataType data_type, DataLayout data_layout) { diff --git a/tests/validation/fixtures/dynamic_fusion/gpu/cl/ElementwiseBinaryFixture.h b/tests/validation/fixtures/dynamic_fusion/gpu/cl/ElementwiseBinaryFixture.h index 22deff1f24..567322f181 100644 --- a/tests/validation/fixtures/dynamic_fusion/gpu/cl/ElementwiseBinaryFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/gpu/cl/ElementwiseBinaryFixture.h @@ -47,7 +47,6 @@ template void setup(ArithmeticOperation ref_op, const TensorShape &shape0, const TensorShape &shape1, const TensorShape &shape2, DataType data_type, bool is_inplace, bool fuse_two_ops = false) { _ref_op = ref_op; @@ -220,7 +219,6 @@ template { public: - template void setup(ArithmeticOperation ref_op, const TensorShape &shape0, DataType data_type, bool is_inplace) { DynamicFusionGpuElementwiseBinaryValidationGenericFixture::setup(ref_op, shape0, shape0, TensorShape(), data_type, is_inplace); @@ -231,7 +229,6 @@ template { public: - template void setup(ArithmeticOperation ref_op, const TensorShape &shape0, const TensorShape &shape1, DataType data_type, bool is_inplace) { DynamicFusionGpuElementwiseBinaryValidationGenericFixture::setup(ref_op, shape0, shape1, TensorShape(), data_type, is_inplace); @@ -242,7 +239,6 @@ template { public: - template void setup(ArithmeticOperation ref_op, const TensorShape &shape0, const TensorShape &shape1, const TensorShape &shape2, DataType data_type, bool is_inplace, bool fuse_two_ops) { DynamicFusionGpuElementwiseBinaryValidationGenericFixture::setup(ref_op, shape0, shape1, shape2, data_type, is_inplace, fuse_two_ops); diff --git a/tests/validation/fixtures/dynamic_fusion/gpu/cl/Pool2dFixture.h b/tests/validation/fixtures/dynamic_fusion/gpu/cl/Pool2dFixture.h index 249f57aceb..0efb761967 100644 --- a/tests/validation/fixtures/dynamic_fusion/gpu/cl/Pool2dFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/gpu/cl/Pool2dFixture.h @@ -51,7 +51,6 @@ template void setup(TensorShape input_shape, const Pool2dAttributes &pool_attr, DataType data_type, bool mixed_precision) { _target = compute_target(input_shape, pool_attr, data_type, mixed_precision); @@ -151,7 +150,6 @@ template { public: - template void setup(TensorShape input_shape, PoolingType pool_type, Size2D pool_size, Padding2D pad, Size2D stride, bool exclude_padding, DataType data_type) { DynamicFusionGpuPool2dValidationGenericFixture::setup(input_shape, @@ -164,7 +162,6 @@ template { public: - template void setup(TensorShape input_shape, PoolingType pool_type, Size2D pool_size, Padding2D pad, Size2D stride, bool exclude_padding, DataType data_type, bool mixed_precision) { DynamicFusionGpuPool2dValidationGenericFixture::setup(input_shape, @@ -177,7 +174,6 @@ template { public: - template void setup(TensorShape input_shape, Pool2dAttributes pool_attr, DataType data_type) { DynamicFusionGpuPool2dValidationGenericFixture::setup(input_shape, pool_attr, data_type, false); diff --git a/tests/validation/fixtures/dynamic_fusion/operators/ActivationFixture.h b/tests/validation/fixtures/dynamic_fusion/operators/ActivationFixture.h index 3fb2cc2b7c..18c3b6bfbb 100644 --- a/tests/validation/fixtures/dynamic_fusion/operators/ActivationFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/operators/ActivationFixture.h @@ -47,7 +47,6 @@ template void setup(TensorShape shape, bool fuse, DataType data_type, ActivationLayerInfo act_info, TArgs... args) { _fuse = fuse; @@ -177,7 +176,6 @@ template { public: - template void setup(TensorShape shape, bool fuse, DataType data_type) { ActivationLayerInfo act_info{ ActivationLayerInfo::ActivationFunction::LOGISTIC }; @@ -189,7 +187,6 @@ template { public: - template void setup(TensorShape shape, bool fuse, DataType data_type) { ActivationLayerInfo act_info{ ActivationLayerInfo::ActivationFunction::TANH }; diff --git a/tests/validation/fixtures/dynamic_fusion/operators/CastFixture.h b/tests/validation/fixtures/dynamic_fusion/operators/CastFixture.h index 8a8e2b0c9a..d8e250cb36 100644 --- a/tests/validation/fixtures/dynamic_fusion/operators/CastFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/operators/CastFixture.h @@ -47,7 +47,6 @@ template void setup(TensorShape shape, DataType dt_in, DataType dt_out, ConvertPolicy policy) { _target = compute_target(shape, dt_in, dt_out, policy); diff --git a/tests/validation/fixtures/dynamic_fusion/operators/ClampFixture.h b/tests/validation/fixtures/dynamic_fusion/operators/ClampFixture.h index cafd28f7b4..3c325d739c 100644 --- a/tests/validation/fixtures/dynamic_fusion/operators/ClampFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/operators/ClampFixture.h @@ -46,7 +46,6 @@ template void setup(TensorShape shape, ClampAttributes attributes, bool fuse, DataType data_type) { // CLAMP is implemented as LU_BOUNDED_RELU with the alpha and beta variables swapped. diff --git a/tests/validation/fixtures/dynamic_fusion/operators/MulFixture.h b/tests/validation/fixtures/dynamic_fusion/operators/MulFixture.h index a0d6bc6ed5..02dc996ffa 100644 --- a/tests/validation/fixtures/dynamic_fusion/operators/MulFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/operators/MulFixture.h @@ -52,7 +52,6 @@ template void setup(const TensorShape &shape0, const TensorShape &shape1, const TensorShape &shape2, DataType data_type, bool is_inplace, bool fuse_two_ops = false) { _data_type = data_type; @@ -200,7 +199,6 @@ template { public: - template void setup(const TensorShape &shape0, DataType data_type, bool is_inplace) { DynamicFusionMulValidationFixture::setup(shape0, shape0, TensorShape(), data_type, is_inplace); @@ -211,7 +209,6 @@ template { public: - template void setup(const TensorShape &shape0, const TensorShape &shape1, DataType data_type, bool is_inplace) { DynamicFusionMulValidationFixture::setup(shape0, shape1, TensorShape(), data_type, is_inplace); @@ -222,7 +219,6 @@ template { public: - template void setup(const TensorShape &shape0, const TensorShape &shape1, const TensorShape &shape2, DataType data_type, bool is_inplace, bool fuse_two_ops) { DynamicFusionMulValidationFixture::setup(shape0, shape1, shape2, data_type, is_inplace, fuse_two_ops); diff --git a/tests/validation/fixtures/dynamic_fusion/operators/ReshapeFixture.h b/tests/validation/fixtures/dynamic_fusion/operators/ReshapeFixture.h index 88c04de35a..abfc6459d6 100644 --- a/tests/validation/fixtures/dynamic_fusion/operators/ReshapeFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/operators/ReshapeFixture.h @@ -50,7 +50,6 @@ template void setup(TensorShape input_shape, TensorShape output_shape, DataType data_type) { _target = compute_target(input_shape, output_shape, data_type); diff --git a/tests/validation/fixtures/dynamic_fusion/operators/ResizeFixture.h b/tests/validation/fixtures/dynamic_fusion/operators/ResizeFixture.h index 62ef053dca..c44f0371d0 100644 --- a/tests/validation/fixtures/dynamic_fusion/operators/ResizeFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/operators/ResizeFixture.h @@ -52,7 +52,6 @@ template void setup(TensorShape shape, DataType data_type, QuantizationInfo quantization_info, DataLayout data_layout, InterpolationPolicy interpolation_policy, SamplingPolicy sampling_policy, bool align_corners, QuantizationInfo output_quantization_info) @@ -223,7 +222,6 @@ template { public: - template void setup(TensorShape shape, DataType data_type, DataLayout data_layout, InterpolationPolicy policy, SamplingPolicy sampling_policy, bool align_corners) { DynamicFusionResizeGenericValidationFixture::setup(shape, @@ -241,7 +239,6 @@ template { public: - template void setup(TensorShape shape, DataType data_type, QuantizationInfo quantization_info, DataLayout data_layout, InterpolationPolicy policy, SamplingPolicy sampling_policy, bool align_corners) { diff --git a/tests/validation/fixtures/dynamic_fusion/operators/SoftmaxFixture.h b/tests/validation/fixtures/dynamic_fusion/operators/SoftmaxFixture.h index 0f50e8e12f..1ed133d2ef 100644 --- a/tests/validation/fixtures/dynamic_fusion/operators/SoftmaxFixture.h +++ b/tests/validation/fixtures/dynamic_fusion/operators/SoftmaxFixture.h @@ -46,7 +46,6 @@ template void setup(TensorShape shape, DataType data_type, float beta, size_t axis, bool is_log) { _reference = compute_reference(shape, data_type, beta, axis, is_log); @@ -143,7 +142,6 @@ template { public: - template void setup(TensorShape shape, DataType data_type, float beta, size_t axis, bool is_log) { DynamicFusionSoftmaxValidationGenericFixture::setup(shape, -- cgit v1.2.1