diff options
Diffstat (limited to 'src/backends/reference/workloads')
9 files changed, 0 insertions, 200 deletions
diff --git a/src/backends/reference/workloads/BaseIterator.hpp b/src/backends/reference/workloads/BaseIterator.hpp index e09371fd96..2d27951b73 100644 --- a/src/backends/reference/workloads/BaseIterator.hpp +++ b/src/backends/reference/workloads/BaseIterator.hpp @@ -260,44 +260,6 @@ private: }; -class BFloat16Decoder : public TypedIterator<const BFloat16, Decoder<float>> -{ -public: - BFloat16Decoder(const BFloat16* data) - : TypedIterator(data) {} - - BFloat16Decoder() - : BFloat16Decoder(nullptr) {} - - float Get() const override - { - float val = 0.f; - armnnUtils::FloatingPointConverter::ConvertBFloat16ToFloat32(m_Iterator, 1, &val); - return val; - } - std::vector<float> DecodeTensor (const TensorShape& tensorShape, - const bool isDepthwise) override - { - IgnoreUnused(isDepthwise); - - const unsigned int size = tensorShape.GetNumElements(); - std::vector<float> decodedTensor; - decodedTensor.reserve(size); - - for (uint32_t i = 0; i < size; ++i) - { - this->operator[](i); - - float val = 0.f; - armnnUtils::FloatingPointConverter::ConvertBFloat16ToFloat32(m_Iterator, 1, &val); - decodedTensor.emplace_back(val); - } - - return decodedTensor; - } - -}; - class Float16Decoder : public TypedIterator<const Half, Decoder<float>> { public: @@ -624,28 +586,6 @@ private: const int32_t m_Offset; }; -class BFloat16Encoder : public TypedIterator<armnn::BFloat16, Encoder<float>> -{ -public: - BFloat16Encoder(armnn::BFloat16* data) - : TypedIterator(data) {} - - BFloat16Encoder() - : BFloat16Encoder(nullptr) {} - - void Set(float right) override - { - armnnUtils::FloatingPointConverter::ConvertFloat32ToBFloat16(&right, 1, m_Iterator); - } - - float Get() const override - { - float val = 0.f; - armnnUtils::FloatingPointConverter::ConvertBFloat16ToFloat32(m_Iterator, 1, &val); - return val; - } -}; - class Float16Encoder : public TypedIterator<Half, Encoder<float>> { public: diff --git a/src/backends/reference/workloads/CMakeLists.txt b/src/backends/reference/workloads/CMakeLists.txt index b8835e3cdb..de6c042959 100644 --- a/src/backends/reference/workloads/CMakeLists.txt +++ b/src/backends/reference/workloads/CMakeLists.txt @@ -88,12 +88,8 @@ list(APPEND armnnRefBackendWorkloads_sources RefConcatWorkload.hpp RefConstantWorkload.cpp RefConstantWorkload.hpp - RefConvertBf16ToFp32Workload.cpp - RefConvertBf16ToFp32Workload.hpp RefConvertFp16ToFp32Workload.cpp RefConvertFp16ToFp32Workload.hpp - RefConvertFp32ToBf16Workload.cpp - RefConvertFp32ToBf16Workload.hpp RefConvertFp32ToFp16Workload.cpp RefConvertFp32ToFp16Workload.hpp RefConvolution2dWorkload.cpp diff --git a/src/backends/reference/workloads/Decoders.hpp b/src/backends/reference/workloads/Decoders.hpp index c2a456bfce..54e7008d50 100644 --- a/src/backends/reference/workloads/Decoders.hpp +++ b/src/backends/reference/workloads/Decoders.hpp @@ -88,10 +88,6 @@ inline std::unique_ptr<Decoder<float>> MakeDecoder(const TensorInfo& info, const info.GetQuantizationScale(), info.GetQuantizationOffset()); } - case DataType::BFloat16: - { - return std::make_unique<BFloat16Decoder>(static_cast<const BFloat16*>(data)); - } case DataType::Float16: { return std::make_unique<Float16Decoder>(static_cast<const Half*>(data)); diff --git a/src/backends/reference/workloads/Encoders.hpp b/src/backends/reference/workloads/Encoders.hpp index a7be9e172b..d6d611494d 100644 --- a/src/backends/reference/workloads/Encoders.hpp +++ b/src/backends/reference/workloads/Encoders.hpp @@ -65,10 +65,6 @@ inline std::unique_ptr<Encoder<float>> MakeEncoder(const TensorInfo& info, void* { return std::make_unique<Int32Encoder>(static_cast<int32_t*>(data)); } - case armnn::DataType::BFloat16: - { - return std::make_unique<BFloat16Encoder>(static_cast<armnn::BFloat16*>(data)); - } case armnn::DataType::Float16: { return std::make_unique<Float16Encoder>(static_cast<Half*>(data)); diff --git a/src/backends/reference/workloads/RefConvertBf16ToFp32Workload.cpp b/src/backends/reference/workloads/RefConvertBf16ToFp32Workload.cpp deleted file mode 100644 index 2fe2eafb9b..0000000000 --- a/src/backends/reference/workloads/RefConvertBf16ToFp32Workload.cpp +++ /dev/null @@ -1,39 +0,0 @@ -// -// Copyright © 2022 Arm Ltd and Contributors. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#include "RefConvertBf16ToFp32Workload.hpp" -#include "RefWorkloadUtils.hpp" - -#include <armnnUtils/FloatingPointConverter.hpp> - -#include <BFloat16.hpp> - -namespace armnn -{ - -void RefConvertBf16ToFp32Workload::Execute() const -{ - Execute(m_Data.m_Inputs, m_Data.m_Outputs); -} - -void RefConvertBf16ToFp32Workload::ExecuteAsync(ExecutionData& executionData) -{ - WorkingMemDescriptor* workingMemDescriptor = static_cast<WorkingMemDescriptor*>(executionData.m_Data); - Execute(workingMemDescriptor->m_Inputs, workingMemDescriptor->m_Outputs); -} - -void RefConvertBf16ToFp32Workload::Execute(std::vector<ITensorHandle*> inputs, - std::vector<ITensorHandle*> outputs) const -{ - ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefConvertBf16ToFp32Workload_Execute"); - - const BFloat16* const input = reinterpret_cast<const BFloat16*>(inputs[0]->Map()); - float* const output = reinterpret_cast<float*>(outputs[0]->Map()); - - unsigned int numElements = GetTensorInfo(inputs[0]).GetNumElements(); - armnnUtils::FloatingPointConverter::ConvertBFloat16ToFloat32(input, numElements, output); -} - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefConvertBf16ToFp32Workload.hpp b/src/backends/reference/workloads/RefConvertBf16ToFp32Workload.hpp deleted file mode 100644 index 24dcb0f682..0000000000 --- a/src/backends/reference/workloads/RefConvertBf16ToFp32Workload.hpp +++ /dev/null @@ -1,24 +0,0 @@ -// -// Copyright © 2022 Arm Ltd and Contributors. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#pragma once - -#include "RefBaseWorkload.hpp" -#include <armnn/backends/WorkloadData.hpp> - -namespace armnn -{ - -class RefConvertBf16ToFp32Workload : public BFloat16ToFloat32Workload<ConvertBf16ToFp32QueueDescriptor> -{ -public: - using BFloat16ToFloat32Workload<ConvertBf16ToFp32QueueDescriptor>::BFloat16ToFloat32Workload; - void Execute() const override; - void ExecuteAsync(ExecutionData& executionData) override; -private: - void Execute(std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs) const; -}; - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefConvertFp32ToBf16Workload.cpp b/src/backends/reference/workloads/RefConvertFp32ToBf16Workload.cpp deleted file mode 100644 index 71ee95b2aa..0000000000 --- a/src/backends/reference/workloads/RefConvertFp32ToBf16Workload.cpp +++ /dev/null @@ -1,39 +0,0 @@ -// -// Copyright © 2022 Arm Ltd and Contributors. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#include "RefConvertFp32ToBf16Workload.hpp" -#include "RefWorkloadUtils.hpp" - -#include <armnnUtils/FloatingPointConverter.hpp> - -#include <BFloat16.hpp> - -namespace armnn -{ - -void RefConvertFp32ToBf16Workload::Execute() const -{ - Execute(m_Data.m_Inputs, m_Data.m_Outputs); -} - -void RefConvertFp32ToBf16Workload::ExecuteAsync(ExecutionData& executionData) -{ - WorkingMemDescriptor* workingMemDescriptor = static_cast<WorkingMemDescriptor*>(executionData.m_Data); - Execute(workingMemDescriptor->m_Inputs, workingMemDescriptor->m_Outputs); -} - -void RefConvertFp32ToBf16Workload::Execute(std::vector<ITensorHandle*> inputs, - std::vector<ITensorHandle*> outputs) const -{ - ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefConvertFp32ToBf16Workload_Execute"); - - const float* const input = reinterpret_cast<const float*>(inputs[0]->Map()); - BFloat16* const output = reinterpret_cast<BFloat16*>(outputs[0]->Map()); - - unsigned int numElements = GetTensorInfo(inputs[0]).GetNumElements(); - armnnUtils::FloatingPointConverter::ConvertFloat32ToBFloat16(input, numElements, output); -} - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefConvertFp32ToBf16Workload.hpp b/src/backends/reference/workloads/RefConvertFp32ToBf16Workload.hpp deleted file mode 100644 index c1e57ec37e..0000000000 --- a/src/backends/reference/workloads/RefConvertFp32ToBf16Workload.hpp +++ /dev/null @@ -1,24 +0,0 @@ -// -// Copyright © 2022 Arm Ltd and Contributors. All rights reserved. -// SPDX-License-Identifier: MIT -// - -#pragma once - -#include "RefBaseWorkload.hpp" -#include <armnn/backends/WorkloadData.hpp> - -namespace armnn -{ - -class RefConvertFp32ToBf16Workload : public Float32ToBFloat16Workload<ConvertFp32ToBf16QueueDescriptor> -{ -public: - using Float32ToBFloat16Workload<ConvertFp32ToBf16QueueDescriptor>::Float32ToBFloat16Workload; - void Execute() const override; - void ExecuteAsync(ExecutionData& executionData) override; -private: - void Execute(std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs) const; -}; - -} //namespace armnn diff --git a/src/backends/reference/workloads/RefWorkloads.hpp b/src/backends/reference/workloads/RefWorkloads.hpp index e049d8db2c..afed71bfff 100644 --- a/src/backends/reference/workloads/RefWorkloads.hpp +++ b/src/backends/reference/workloads/RefWorkloads.hpp @@ -17,9 +17,7 @@ #include "RefConvolution3dWorkload.hpp" #include "RefConstantWorkload.hpp" #include "RefConcatWorkload.hpp" -#include "RefConvertBf16ToFp32Workload.hpp" #include "RefConvertFp16ToFp32Workload.hpp" -#include "RefConvertFp32ToBf16Workload.hpp" #include "RefConvertFp32ToFp16Workload.hpp" #include "RefDebugWorkload.hpp" #include "RefDepthToSpaceWorkload.hpp" |