13 using namespace armcomputetensorutils;
15 static constexpr arm_compute::ConvertPolicy g_AclConvertPolicy = arm_compute::ConvertPolicy::SATURATE;
20 const arm_compute::CLCompileContext& clCompileContext) :
30 m_Layer.configure(clCompileContext, &input, &output, g_AclConvertPolicy, 0);
44 return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR,
"Input should be Float32");
48 return arm_compute::Status(arm_compute::ErrorCode::RUNTIME_ERROR,
"Output should be Float16");
51 const arm_compute::TensorInfo aclInputInfo = BuildArmComputeTensorInfo(input);
52 const arm_compute::TensorInfo aclOutputInfo = BuildArmComputeTensorInfo(output);
55 &aclInputInfo, &aclOutputInfo, g_AclConvertPolicy, 0);
#define ARMNN_SCOPED_PROFILING_EVENT_CL_GUID(name, guid)
arm_compute::Status ClConvertFp32ToFp16WorkloadValidate(const TensorInfo &input, const TensorInfo &output)
void RunClFunction(arm_compute::IFunction &function, const CheckLocation &location)
void ValidateInputsOutputs(const std::string &descName, unsigned int numExpectedIn, unsigned int numExpectedOut) const
Copyright (c) 2021 ARM Limited and Contributors.
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
DataType GetDataType() const
ClConvertFp32ToFp16Workload(const ConvertFp32ToFp16QueueDescriptor &descriptor, const WorkloadInfo &info, const arm_compute::CLCompileContext &clCompileContext)
virtual void Execute() const override
profiling::ProfilingGuid GetGuid() const final
std::vector< ITensorHandle * > m_Outputs
Contains information about TensorInfos of a layer.
std::vector< ITensorHandle * > m_Inputs