aboutsummaryrefslogtreecommitdiff
path: root/src/backends/reference/workloads/Encoders.hpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/backends/reference/workloads/Encoders.hpp')
-rw-r--r--src/backends/reference/workloads/Encoders.hpp113
1 files changed, 0 insertions, 113 deletions
diff --git a/src/backends/reference/workloads/Encoders.hpp b/src/backends/reference/workloads/Encoders.hpp
deleted file mode 100644
index f52297602f..0000000000
--- a/src/backends/reference/workloads/Encoders.hpp
+++ /dev/null
@@ -1,113 +0,0 @@
-//
-// Copyright © 2017 Arm Ltd. All rights reserved.
-// SPDX-License-Identifier: MIT
-//
-
-#pragma once
-
-#include "BaseIterator.hpp"
-
-#include <armnnUtils/TensorUtils.hpp>
-
-#include <boost/assert.hpp>
-
-namespace armnn
-{
-
-template<typename T>
-inline std::unique_ptr<Encoder<T>> MakeEncoder(const TensorInfo& info, void* data = nullptr);
-
-template<>
-inline std::unique_ptr<Encoder<float>> MakeEncoder(const TensorInfo& info, void* data)
-{
- switch(info.GetDataType())
- {
- ARMNN_NO_DEPRECATE_WARN_BEGIN
- case armnn::DataType::QuantizedSymm8PerAxis:
- {
- std::pair<unsigned int, std::vector<float>> params = armnnUtils::GetPerAxisParams(info);
- return std::make_unique<QSymm8PerAxisEncoder>(
- static_cast<int8_t*>(data),
- params.second,
- params.first);
- }
- ARMNN_NO_DEPRECATE_WARN_END
- case armnn::DataType::QAsymmS8:
- {
- return std::make_unique<QASymmS8Encoder>(
- static_cast<int8_t*>(data),
- info.GetQuantizationScale(),
- info.GetQuantizationOffset());
- }
- case armnn::DataType::QAsymmU8:
- {
- return std::make_unique<QASymm8Encoder>(
- static_cast<uint8_t*>(data),
- info.GetQuantizationScale(),
- info.GetQuantizationOffset());
- }
- case DataType::QSymmS8:
- {
- if (info.HasPerAxisQuantization())
- {
- std::pair<unsigned int, std::vector<float>> params = armnnUtils::GetPerAxisParams(info);
- return std::make_unique<QSymm8PerAxisEncoder>(
- static_cast<int8_t*>(data),
- params.second,
- params.first);
- }
- else
- {
- return std::make_unique<QSymmS8Encoder>(
- static_cast<int8_t*>(data),
- info.GetQuantizationScale(),
- info.GetQuantizationOffset());
- }
- }
- case armnn::DataType::QSymmS16:
- {
- return std::make_unique<QSymm16Encoder>(
- static_cast<int16_t*>(data),
- info.GetQuantizationScale(),
- info.GetQuantizationOffset());
- }
- case armnn::DataType::Signed32:
- {
- return std::make_unique<Int32Encoder>(static_cast<int32_t*>(data));
- }
- case armnn::DataType::Float16:
- {
- return std::make_unique<Float16Encoder>(static_cast<Half*>(data));
- }
- case armnn::DataType::Float32:
- {
- return std::make_unique<Float32Encoder>(static_cast<float*>(data));
- }
- default:
- {
- BOOST_ASSERT_MSG(false, "Unsupported target Data Type!");
- break;
- }
- }
- return nullptr;
-}
-
-template<>
-inline std::unique_ptr<Encoder<bool>> MakeEncoder(const TensorInfo& info, void* data)
-{
- switch(info.GetDataType())
- {
- case armnn::DataType::Boolean:
- {
- return std::make_unique<BooleanEncoder>(static_cast<uint8_t*>(data));
- }
- default:
- {
- BOOST_ASSERT_MSG(false, "Cannot encode from boolean. Not supported target Data Type!");
- break;
- }
- }
- return nullptr;
-}
-
-} //namespace armnn