From 34336f9378eca4d39913ac0d9ba411a494631ad5 Mon Sep 17 00:00:00 2001 From: Matthew Bentham Date: Thu, 27 Apr 2023 12:13:50 +0000 Subject: Make Convert workloads use arm_compute::NECast in CpuAcc backend NECast can use conversion instructions where they are available so this should in general be faster. Signed-off-by: Matthew Bentham Change-Id: I3f259e17b280a4f4c36f363965ffbc8ee8c4c29f --- src/backends/neon/workloads/NeonConvertFp32ToFp16Workload.hpp | 7 ++++++- 1 file changed, 6 insertions(+), 1 deletion(-) (limited to 'src/backends/neon/workloads/NeonConvertFp32ToFp16Workload.hpp') diff --git a/src/backends/neon/workloads/NeonConvertFp32ToFp16Workload.hpp b/src/backends/neon/workloads/NeonConvertFp32ToFp16Workload.hpp index 666f48794b..c6fed76e6d 100644 --- a/src/backends/neon/workloads/NeonConvertFp32ToFp16Workload.hpp +++ b/src/backends/neon/workloads/NeonConvertFp32ToFp16Workload.hpp @@ -5,13 +5,17 @@ #pragma once +#include #include #include +#include #include namespace armnn { +arm_compute::Status NeonConvertFp32ToFp16WorkloadValidate(const TensorInfo& input, const TensorInfo& output); + class NeonConvertFp32ToFp16Workload : public Float32ToFloat16Workload { public: @@ -23,9 +27,10 @@ public: // Replace output tensor handle with the given TensorHandle void ReplaceOutputTensorHandle(ITensorHandle* tensorHandle, unsigned int slot) override; private: + virtual void Reconfigure(); using TensorHandlePair = std::pair; std::vector m_TensorHandlePairs; - virtual void Reconfigure(); + mutable std::unique_ptr m_Cast; }; } //namespace armnn -- cgit v1.2.1