20 const arm_compute::TensorInfo neonOutputInfo = armcomputetensorutils::BuildArmComputeTensorInfo(output);
22 std::array<arm_compute::DataType,8> supportedTypes = {
23 arm_compute::DataType::F16,
24 arm_compute::DataType::F32,
25 arm_compute::DataType::QASYMM8,
26 arm_compute::DataType::QASYMM8_SIGNED,
27 arm_compute::DataType::QSYMM16,
28 arm_compute::DataType::QSYMM8,
29 arm_compute::DataType::QSYMM8_PER_CHANNEL,
30 arm_compute::DataType::S32
32 auto it = std::find(begin(supportedTypes), end(supportedTypes), neonOutputInfo.data_type());
34 if (it != end(supportedTypes))
66 switch (computeDataType)
68 case arm_compute::DataType::F16:
73 case arm_compute::DataType::F32:
78 case arm_compute::DataType::QASYMM8:
83 case arm_compute::DataType::QASYMM8_SIGNED:
88 case arm_compute::DataType::QSYMM16:
93 case arm_compute::DataType::QSYMM8:
94 case arm_compute::DataType::QSYMM8_PER_CHANNEL:
99 case arm_compute::DataType::S32:
void Execute() const override
#define ARMNN_SCOPED_PROFILING_EVENT_CL(name)
const ConstantQueueDescriptor m_Data
const ConstCpuTensorHandle * m_LayerOutput
const T * GetConstTensor() const
Copyright (c) 2020 ARM Limited.
#define ARMNN_ASSERT_MSG(COND, MSG)
arm_compute::Status ClConstantWorkloadValidate(const TensorInfo &output)
#define ARMNN_ASSERT(COND)
ClConstantWorkload(const ConstantQueueDescriptor &descriptor, const WorkloadInfo &info)
void CopyArmComputeClTensorData(arm_compute::CLTensor &dstTensor, const T *srcData)
std::vector< ITensorHandle * > m_Outputs
Contains information about inputs and outputs to a layer.