// // Copyright © 2017 Arm Ltd. All rights reserved. // See LICENSE file in the project root for full license information. // #include "ClDivisionFloatWorkload.hpp" #include "backends/ClTensorHandle.hpp" #include "backends/CpuTensorHandle.hpp" namespace armnn { arm_compute::Status ClDivisionWorkloadValidate(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output) { const arm_compute::TensorInfo aclInput1 = armcomputetensorutils::BuildArmComputeTensorInfo(input0); const arm_compute::TensorInfo aclInput2 = armcomputetensorutils::BuildArmComputeTensorInfo(input1); const arm_compute::TensorInfo aclOutput = armcomputetensorutils::BuildArmComputeTensorInfo(output); // At the time of writing, configure() will fail if a rounding policy other than TO_ZERO is supplied to it, // when providing a scale of 1.0 for F32 tensors, even though the provided rounding policy appears to be // ignored for F32 tensors. return arm_compute::CLArithmeticDivision::validate(&aclInput1, &aclInput2, &aclOutput); } ClDivisionFloatWorkload::ClDivisionFloatWorkload(const DivisionQueueDescriptor& descriptor, const WorkloadInfo& info) : FloatWorkload(descriptor, info) { m_Data.ValidateInputsOutputs("ClDivisionFloatWorkload", 2, 1); arm_compute::ICLTensor& input0 = static_cast(m_Data.m_Inputs[0])->GetTensor(); arm_compute::ICLTensor& input1 = static_cast(m_Data.m_Inputs[1])->GetTensor(); arm_compute::ICLTensor& output = static_cast(m_Data.m_Outputs[0])->GetTensor(); // Construct m_ArithmeticDivision.configure(&input0, &input1, &output); } void ClDivisionFloatWorkload::Execute() const { ARMNN_SCOPED_PROFILING_EVENT_CL("ClDivisionFloatWorkload_Execute"); // Executes the layer. m_ArithmeticDivision.run(); } } //namespace armnn