aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMichele Di Giorgio <michele.digiorgio@arm.com>2021-04-28 16:03:32 +0100
committerMichele Di Giorgio <michele.digiorgio@arm.com>2021-04-29 16:19:20 +0000
commit223aba9ff116ecc7db7963ddb6a7213eee04b18c (patch)
tree0528c22e74b25f9fd1fd41e8f29651b0fce09102
parent9f6111a9e0746d5995fd9dc46420af5b316be734 (diff)
downloadComputeLibrary-223aba9ff116ecc7db7963ddb6a7213eee04b18c.tar.gz
Remove stale/solved TODOs
Change-Id: I5c440f4c6ca4186adcfa926e6b7d924086671f29 Signed-off-by: Michele Di Giorgio <michele.digiorgio@arm.com> Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/5520 Tested-by: Arm Jenkins <bsgcomp@arm.com> Reviewed-by: Manuel Bottini <manuel.bottini@arm.com> Comments-Addressed: Arm Jenkins <bsgcomp@arm.com>
-rw-r--r--src/core/CL/cl_kernels/softmax_layer.cl4
-rw-r--r--src/core/CL/cl_kernels/softmax_layer_quantized.cl4
-rw-r--r--src/core/NEON/NEMath.inl3
-rw-r--r--src/graph/GraphManager.cpp2
-rw-r--r--src/graph/backends/CL/CLFunctionsFactory.cpp1
-rw-r--r--src/runtime/NEON/functions/NEGEMMConvolutionLayer.cpp2
-rw-r--r--tests/datasets/SmallConvolutionLayerDataset.h3
-rw-r--r--tests/validation/CL/NormalizationLayer.cpp2
-rw-r--r--tests/validation/NEON/GEMM.cpp4
-rw-r--r--tests/validation/fixtures/GEMMLowpFixture.h1
10 files changed, 4 insertions, 22 deletions
diff --git a/src/core/CL/cl_kernels/softmax_layer.cl b/src/core/CL/cl_kernels/softmax_layer.cl
index 01f5de47cf..4d2d89dd73 100644
--- a/src/core/CL/cl_kernels/softmax_layer.cl
+++ b/src/core/CL/cl_kernels/softmax_layer.cl
@@ -1,5 +1,5 @@
/*
- * Copyright (c) 2017-2020 Arm Limited.
+ * Copyright (c) 2017-2021 Arm Limited.
*
* SPDX-License-Identifier: MIT
*
@@ -306,7 +306,6 @@ __kernel void softmax_layer_max_shift_exp_sum_parallel(
}
#ifdef NON_MULTIPLE_OF_GRID_SIZE
// How many work-items needed to complete the computation.
- //TODO: Optimize this calculation (avoid %).
int boundary_workitems = (SRC_WIDTH % (GRID_SIZE * VECTOR_SIZE)) / VECTOR_SIZE;
if(lid < boundary_workitems)
{
@@ -417,7 +416,6 @@ __kernel void softmax_layer_max_shift_exp_sum_parallel(
sum1D += data;
}
#ifdef NON_MULTIPLE_OF_GRID_SIZE
- //TODO: Optimize the calculation (avoid %).
boundary_workitems = (SRC_WIDTH % (GRID_SIZE * VECTOR_SIZE)) / VECTOR_SIZE;
if(lid < boundary_workitems)
{
diff --git a/src/core/CL/cl_kernels/softmax_layer_quantized.cl b/src/core/CL/cl_kernels/softmax_layer_quantized.cl
index b7a6e00dfa..4d5006d804 100644
--- a/src/core/CL/cl_kernels/softmax_layer_quantized.cl
+++ b/src/core/CL/cl_kernels/softmax_layer_quantized.cl
@@ -1,5 +1,5 @@
/*
- * Copyright (c) 2017-2020 Arm Limited.
+ * Copyright (c) 2017-2021 Arm Limited.
*
* SPDX-License-Identifier: MIT
*
@@ -324,7 +324,6 @@ __kernel void softmax_layer_max_shift_exp_sum_quantized_parallel(
}
#ifdef NON_MULTIPLE_OF_GRID_SIZE
// How many work-items needed to complete the computation.
- //TODO: Optimize this calculation (avoid %).
int boundary_workitems = (SRC_WIDTH % (GRID_SIZE * VECTOR_SIZE)) / VECTOR_SIZE;
if(lid < boundary_workitems)
{
@@ -429,7 +428,6 @@ __kernel void softmax_layer_max_shift_exp_sum_quantized_parallel(
sum1D = sum1D + select(0, data_fp, data_diff >= (VEC_INT)(DIFF_MIN));
}
#ifdef NON_MULTIPLE_OF_GRID_SIZE
- //TODO: Optimize the calculation (avoid %).
boundary_workitems = (SRC_WIDTH % (GRID_SIZE * VECTOR_SIZE)) / VECTOR_SIZE;
if(lid < boundary_workitems)
{
diff --git a/src/core/NEON/NEMath.inl b/src/core/NEON/NEMath.inl
index da9d038139..29df5433bb 100644
--- a/src/core/NEON/NEMath.inl
+++ b/src/core/NEON/NEMath.inl
@@ -495,7 +495,6 @@ inline float16x8_t vtaylor_polyq_f16(float16x8_t x, const std::array<float16x8_t
inline float16x8_t vexpq_f16(float16x8_t x)
{
- // TODO (COMPMID-1535) : Revisit FP16 approximations
const float32x4_t x_high = vcvt_f32_f16(vget_high_f16(x));
const float32x4_t x_low = vcvt_f32_f16(vget_low_f16(x));
@@ -505,7 +504,6 @@ inline float16x8_t vexpq_f16(float16x8_t x)
inline float16x8_t vlogq_f16(float16x8_t x)
{
- // TODO (COMPMID-1535) : Revisit FP16 approximations
const float32x4_t x_high = vcvt_f32_f16(vget_high_f16(x));
const float32x4_t x_low = vcvt_f32_f16(vget_low_f16(x));
@@ -515,7 +513,6 @@ inline float16x8_t vlogq_f16(float16x8_t x)
inline float16x8_t vpowq_f16(float16x8_t val, float16x8_t n)
{
- // TODO (giaiod01) - COMPMID-1535
float32x4_t n0_f32 = vcvt_f32_f16(vget_low_f16(n));
float32x4_t n1_f32 = vcvt_f32_f16(vget_high_f16(n));
float32x4_t val0_f32 = vcvt_f32_f16(vget_low_f16(val));
diff --git a/src/graph/GraphManager.cpp b/src/graph/GraphManager.cpp
index ab7aac6230..e357f10401 100644
--- a/src/graph/GraphManager.cpp
+++ b/src/graph/GraphManager.cpp
@@ -55,7 +55,6 @@ void GraphManager::finalize_graph(Graph &graph, GraphContext &ctx, PassManager &
pm.run_type(graph, IGraphMutator::MutationType::IR);
// Force target to all graph construct
- // TODO (COMPMID-2014) : Support heterogeneous execution
Target forced_target = target;
// In case CLVK is selected, use the CL backend and
@@ -77,7 +76,6 @@ void GraphManager::finalize_graph(Graph &graph, GraphContext &ctx, PassManager &
force_target_to_graph(graph, forced_target);
// Setup backend context
- // TODO (COMPMID-2014) : Setup all backends needed by the graph
setup_requested_backend_context(ctx, forced_target);
// Configure all tensors
diff --git a/src/graph/backends/CL/CLFunctionsFactory.cpp b/src/graph/backends/CL/CLFunctionsFactory.cpp
index a3e7261721..1cd3f9f9c7 100644
--- a/src/graph/backends/CL/CLFunctionsFactory.cpp
+++ b/src/graph/backends/CL/CLFunctionsFactory.cpp
@@ -83,7 +83,6 @@ struct CLFusedLayerTypes
using FuseBatchNormalization = CLFuseBatchNormalization;
};
-// TODO (isagot01): Remove once we support heterogeneous scheduling at function level
/** Wrapper for the CPP Function in the OpenCL backend **/
class CPPWrapperFunction : public IFunction
{
diff --git a/src/runtime/NEON/functions/NEGEMMConvolutionLayer.cpp b/src/runtime/NEON/functions/NEGEMMConvolutionLayer.cpp
index 900330e4a6..2876c254fa 100644
--- a/src/runtime/NEON/functions/NEGEMMConvolutionLayer.cpp
+++ b/src/runtime/NEON/functions/NEGEMMConvolutionLayer.cpp
@@ -524,7 +524,6 @@ Status NEGEMMConvolutionLayer::validate(const ITensorInfo *input, const ITensorI
{
// Create tensor info for im2col reshaped inputs
// For CPU, the batch size is on the fourth dimension
- // TODO (giaiod01): Auto-initialize the output shape of im2col COMPMID-1482
TensorShape shape_im2col = input->tensor_shape();
shape_im2col.set(0, mat_weights_rows);
shape_im2col.set(1, conv_w * conv_h);
@@ -532,7 +531,6 @@ Status NEGEMMConvolutionLayer::validate(const ITensorInfo *input, const ITensorI
im2col_reshaped_info = TensorInfo(shape_im2col, 1, data_type);
im2col_reshaped_info.set_quantization_info(input->quantization_info());
-
ARM_COMPUTE_RETURN_ON_ERROR(NEIm2ColKernel::validate(input, &im2col_reshaped_info, Size2D(kernel_width, kernel_height), conv_info, append_bias, dilation));
gemm_input_to_use = &im2col_reshaped_info;
}
diff --git a/tests/datasets/SmallConvolutionLayerDataset.h b/tests/datasets/SmallConvolutionLayerDataset.h
index 66640dd943..7d1db5a73e 100644
--- a/tests/datasets/SmallConvolutionLayerDataset.h
+++ b/tests/datasets/SmallConvolutionLayerDataset.h
@@ -1,5 +1,5 @@
/*
- * Copyright (c) 2017-2020 Arm Limited.
+ * Copyright (c) 2017-2021 Arm Limited.
*
* SPDX-License-Identifier: MIT
*
@@ -181,7 +181,6 @@ public:
}
};
-// TODO (COMPMID-1749)
class SmallConvolutionLayerReducedDataset final : public ConvolutionLayerDataset
{
public:
diff --git a/tests/validation/CL/NormalizationLayer.cpp b/tests/validation/CL/NormalizationLayer.cpp
index b1c28ad644..8c1890842b 100644
--- a/tests/validation/CL/NormalizationLayer.cpp
+++ b/tests/validation/CL/NormalizationLayer.cpp
@@ -62,8 +62,6 @@ const auto NormalizationDatasetFP32 = combine(combine(combine(framework::dataset
TEST_SUITE(CL)
TEST_SUITE(NormalizationLayer)
-//TODO(COMPMID-415): Missing configuration?
-
// *INDENT-OFF*
// clang-format off
DATA_TEST_CASE(Validate, framework::DatasetMode::ALL, zip(zip(zip(
diff --git a/tests/validation/NEON/GEMM.cpp b/tests/validation/NEON/GEMM.cpp
index 2d8c61164b..500c6029d5 100644
--- a/tests/validation/NEON/GEMM.cpp
+++ b/tests/validation/NEON/GEMM.cpp
@@ -1,5 +1,5 @@
/*
- * Copyright (c) 2017-2020 Arm Limited.
+ * Copyright (c) 2017-2021 Arm Limited.
*
* SPDX-License-Identifier: MIT
*
@@ -188,8 +188,6 @@ TEST_SUITE_END() // U8
TEST_SUITE_END() // INTERLEAVE_4X4
-//TODO(COMPMID-415): Validate valid region
-
template <typename T>
using NEGEMMFixture = GEMMValidationFixture<Tensor, Accessor, NEGEMM, T>;
diff --git a/tests/validation/fixtures/GEMMLowpFixture.h b/tests/validation/fixtures/GEMMLowpFixture.h
index c3da2e261d..9e5eb53381 100644
--- a/tests/validation/fixtures/GEMMLowpFixture.h
+++ b/tests/validation/fixtures/GEMMLowpFixture.h
@@ -126,7 +126,6 @@ TensorType compute_gemmlowp_target(const TensorShape &shape_a, const TensorShape
// Create and configure function
// The GEMMinfo includes the values of the depth in case of reinterpreted 3d input/output
FunctionType gemmlowp;
- // TODO (COMPMID-1672) - Extending the test to validate add bias in offset contribution
gemmlowp.configure(&a, &b, is_fused ? &bias : nullptr, &output, GEMMInfo(false, false, false, (reinterpret_output_as_3d ? shape_output[2] : 0), reinterpret_input_as_3d, false, output_stage));
ARM_COMPUTE_EXPECT(a.info()->is_resizable(), framework::LogLevel::ERRORS);