aboutsummaryrefslogtreecommitdiff
path: root/compute_kernel_writer/validation/tests/TensorBitMaskTest.h
diff options
context:
space:
mode:
authorGian Marco Iodice <gianmarco.iodice@arm.com>2023-07-07 11:25:57 +0100
committerViet-Hoa Do <viet-hoa.do@arm.com>2023-07-20 08:48:15 +0000
commitebfdb5a1ea73c2269eec5af492970c2174ab7d0f (patch)
tree5c60d083100118f0a40e629dfa69b7a7373dd7fd /compute_kernel_writer/validation/tests/TensorBitMaskTest.h
parent29dc9fc1d3d6e90746ba1173e3318b774dcf7bed (diff)
downloadComputeLibrary-ebfdb5a1ea73c2269eec5af492970c2174ab7d0f.tar.gz
Integrate CLTensorArgument
- Add CLTensorArgument to query the components and storages as OpenCL variables (or by values when possible) - Add caching mechanism in CLTensorArgument to keep track of the components and storages used - Add unit tests Resolves COMPMID-5787 Signed-off-by: Gian Marco Iodice <gianmarco.iodice@arm.com> Signed-off-by: Viet-Hoa Do <viet-hoa.do@arm.com> Change-Id: Ib39e1f77b097e5b907a296fe6b0d41bb4bcd4ffc Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/9908 Comments-Addressed: Arm Jenkins <bsgcomp@arm.com> Benchmark: Arm Jenkins <bsgcomp@arm.com> Tested-by: Arm Jenkins <bsgcomp@arm.com> Reviewed-by: Gunes Bayir <gunes.bayir@arm.com> Reviewed-by: Jakub Sujak <jakub.sujak@arm.com>
Diffstat (limited to 'compute_kernel_writer/validation/tests/TensorBitMaskTest.h')
-rw-r--r--compute_kernel_writer/validation/tests/TensorBitMaskTest.h221
1 files changed, 221 insertions, 0 deletions
diff --git a/compute_kernel_writer/validation/tests/TensorBitMaskTest.h b/compute_kernel_writer/validation/tests/TensorBitMaskTest.h
new file mode 100644
index 0000000000..759d926d18
--- /dev/null
+++ b/compute_kernel_writer/validation/tests/TensorBitMaskTest.h
@@ -0,0 +1,221 @@
+/*
+ * Copyright (c) 2023 Arm Limited.
+ *
+ * SPDX-License-Identifier: MIT
+ *
+ * Permission is hereby granted, free of charge, to any person obtaining a copy
+ * of this software and associated documentation files (the "Software"), to
+ * deal in the Software without restriction, including without limitation the
+ * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
+ * sell copies of the Software, and to permit persons to whom the Software is
+ * furnished to do so, subject to the following conditions:
+ *
+ * The above copyright notice and this permission notice shall be included in all
+ * copies or substantial portions of the Software.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
+ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
+ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
+ * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
+ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
+ * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
+ * SOFTWARE.
+ */
+#ifndef CKW_TESTS_TENSORBITMASKTEST_H
+#define CKW_TESTS_TENSORBITMASKTEST_H
+
+#include "ckw/TensorInfo.h"
+#include "ckw/types/TensorComponentType.h"
+#include "common/Common.h"
+#include "src/types/TensorComponentType.h"
+
+#include <vector>
+
+namespace ckw
+{
+class TensorBitMaskTrueTest : public ITest
+{
+public:
+ TensorBitMaskTrueTest()
+ {
+ _component.push_back(TensorComponentType::Dim0);
+ _component.push_back(TensorComponentType::Dim1);
+ _component.push_back(TensorComponentType::Dim2);
+ _component.push_back(TensorComponentType::Dim3);
+ _component.push_back(TensorComponentType::Dim4);
+ _component.push_back(TensorComponentType::Stride0);
+ _component.push_back(TensorComponentType::Stride1);
+ _component.push_back(TensorComponentType::Stride2);
+ _component.push_back(TensorComponentType::Stride3);
+ _component.push_back(TensorComponentType::Stride4);
+ _component.push_back(TensorComponentType::Dim1xDim2);
+ _component.push_back(TensorComponentType::Dim1xDim2xDim3);
+ _component.push_back(TensorComponentType::Dim2xDim3);
+ _component.push_back(TensorComponentType::OffsetFirstElement);
+
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ }
+
+ bool run() override
+ {
+ // The status of this variable can change in VALIDATE_TEST()
+ bool all_tests_passed = true;
+
+ VALIDATE_ON_MSG(_component.size() == _bitmask.size(),
+ "The number of layouts and components does not match");
+ const size_t num_tests = _component.size();
+ for(size_t i = 0; i < num_tests; ++i)
+ {
+ const TensorComponentType component = _component[i];
+ const TensorComponentBitmask bitmask = _bitmask[i];
+ const bool out = static_cast<uint32_t>(component) & static_cast<uint32_t>(bitmask);
+ VALIDATE_TEST(out == true, all_tests_passed, i);
+ }
+ return all_tests_passed;
+ }
+
+ std::string name() override
+ {
+ return "TensorBitMaskTrueTest";
+ }
+
+private:
+ std::vector<TensorComponentType> _component{};
+ std::vector<TensorComponentBitmask> _bitmask{};
+};
+
+class TensorBitMaskFalseTest : public ITest
+{
+public:
+ TensorBitMaskFalseTest()
+ {
+ _component.push_back(TensorComponentType::Dim0);
+ _component.push_back(TensorComponentType::Dim1);
+ _component.push_back(TensorComponentType::Dim2);
+ _component.push_back(TensorComponentType::Dim3);
+ _component.push_back(TensorComponentType::Dim4);
+ _component.push_back(TensorComponentType::Dim0);
+ _component.push_back(TensorComponentType::Dim1);
+ _component.push_back(TensorComponentType::Dim2);
+ _component.push_back(TensorComponentType::Dim3);
+ _component.push_back(TensorComponentType::Dim4);
+ _component.push_back(TensorComponentType::Dim0);
+ _component.push_back(TensorComponentType::Dim1);
+ _component.push_back(TensorComponentType::Dim2);
+ _component.push_back(TensorComponentType::Dim3);
+ _component.push_back(TensorComponentType::Dim4);
+ _component.push_back(TensorComponentType::Stride0);
+ _component.push_back(TensorComponentType::Stride1);
+ _component.push_back(TensorComponentType::Stride2);
+ _component.push_back(TensorComponentType::Stride3);
+ _component.push_back(TensorComponentType::Stride4);
+ _component.push_back(TensorComponentType::Stride0);
+ _component.push_back(TensorComponentType::Stride1);
+ _component.push_back(TensorComponentType::Stride2);
+ _component.push_back(TensorComponentType::Stride3);
+ _component.push_back(TensorComponentType::Stride4);
+ _component.push_back(TensorComponentType::Stride0);
+ _component.push_back(TensorComponentType::Stride1);
+ _component.push_back(TensorComponentType::Stride2);
+ _component.push_back(TensorComponentType::Stride3);
+ _component.push_back(TensorComponentType::Stride4);
+ _component.push_back(TensorComponentType::Dim1xDim2);
+ _component.push_back(TensorComponentType::Dim1xDim2xDim3);
+ _component.push_back(TensorComponentType::Dim2xDim3);
+ _component.push_back(TensorComponentType::Dim1xDim2);
+ _component.push_back(TensorComponentType::Dim1xDim2xDim3);
+ _component.push_back(TensorComponentType::Dim2xDim3);
+ _component.push_back(TensorComponentType::Dim1xDim2);
+ _component.push_back(TensorComponentType::Dim1xDim2xDim3);
+ _component.push_back(TensorComponentType::Dim2xDim3);
+ _component.push_back(TensorComponentType::OffsetFirstElement);
+ _component.push_back(TensorComponentType::OffsetFirstElement);
+ _component.push_back(TensorComponentType::OffsetFirstElement);
+
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::OffsetFirstElement);
+ _bitmask.push_back(TensorComponentBitmask::Dimension);
+ _bitmask.push_back(TensorComponentBitmask::Stride);
+ _bitmask.push_back(TensorComponentBitmask::FoldedDimensions);
+ }
+
+ bool run() override
+ {
+ // The status of this variable can change in VALIDATE_TEST()
+ bool all_tests_passed = true;
+
+ VALIDATE_ON_MSG(_component.size() == _bitmask.size(),
+ "The number of layouts and components does not match");
+ const size_t num_tests = _component.size();
+ for(size_t i = 0; i < num_tests; ++i)
+ {
+ const TensorComponentType component = _component[i];
+ const TensorComponentBitmask bitmask = _bitmask[i];
+ const bool out = static_cast<uint32_t>(component) & static_cast<uint32_t>(bitmask);
+ VALIDATE_TEST(out == false, all_tests_passed, i);
+ }
+ return all_tests_passed;
+ }
+
+ std::string name() override
+ {
+ return "TensorBitMaskFalseTest";
+ }
+
+private:
+ std::vector<TensorComponentType> _component{};
+ std::vector<TensorComponentBitmask> _bitmask{};
+};
+} // namespace ckw
+
+#endif // CKW_TESTS_TENSORBITMASKTEST_H