26 #include <arm_compute/core/CL/CLKernelLibrary.h> 27 #include <arm_compute/runtime/CL/CLBufferAllocator.h> 28 #include <arm_compute/runtime/CL/CLScheduler.h> 40 std::string& outReasonIfUnsupported)
47 std::string& outReasonIfUnsupported,
59 std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(
const QueueDescriptorType& descriptor,
65 return MakeWorkloadHelper<FloatWorkload, Uint8Workload>(descriptor,
info, std::forward<Args>(args)...);
73 template <
typename Workload,
typename QueueDescriptorType,
typename... Args>
74 std::unique_ptr<IWorkload> ClWorkloadFactory::MakeWorkload(
const QueueDescriptorType& descriptor,
80 return std::make_unique<Workload>(descriptor,
info, std::forward<Args>(args)...);
89 : m_MemoryManager(memoryManager), m_ModelContextPtr(
IBackendInternal::IBackendSpecificModelContextPtr{})
95 : m_MemoryManager(memoryManager), m_ModelContextPtr(modelContextPtr)
100 const bool IsMemoryManaged)
const 103 std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo);
104 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
111 const bool IsMemoryManaged)
const 114 std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo, dataLayout);
115 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
122 unsigned int const* subTensorOrigin)
const 125 arm_compute::TensorShape shape = armcomputetensorutils::BuildArmComputeTensorShape(subTensorShape);
132 coords.set(i, armnn::numeric_cast<int>(subTensorOrigin[revertedIndex]));
135 const arm_compute::TensorShape parentShape = armcomputetensorutils::BuildArmComputeTensorShape(parent.
GetShape());
136 if (!::arm_compute::error_on_invalid_subtensor(__func__, __FILE__, __LINE__, parentShape, coords, shape))
141 return std::make_unique<ClSubTensorHandle>(
142 PolymorphicDowncast<IClTensorHandle*>(&parent), shape, coords);
159 return MakeWorkload<ClActivationWorkload>(descriptor,
info);
165 return MakeWorkload<ClAdditionWorkload>(descriptor,
info);
171 return std::make_unique<ClArgMinMaxWorkload>(descriptor,
info);
178 return MakeWorkload<ClBatchNormalizationFloatWorkload, NullWorkload>(descriptor,
info);
184 return MakeWorkload<ClBatchToSpaceNdWorkload>(descriptor,
info);
190 return MakeWorkload<ClComparisonWorkload>(descriptor,
info);
196 return MakeWorkload<ClConcatWorkload>(descriptor,
info);
202 return MakeWorkload<ClConstantWorkload>(descriptor,
info);
209 return MakeWorkload<ClConvertFp16ToFp32Workload>(descriptor,
info);
216 return MakeWorkload<ClConvertFp32ToFp16Workload>(descriptor,
info);
222 bool isFastMathEnabled =
false;
223 if (m_ModelContextPtr)
225 if (m_ModelContextPtr.get() !=
nullptr)
234 return MakeWorkload<ClConvolution2dWorkload>(descriptor,
236 m_MemoryManager->GetIntraLayerManager(),
243 return MakeWorkload<NullWorkload, NullWorkload>(descriptor,
info);
249 return MakeWorkload<ClDepthToSpaceWorkload>(descriptor,
info);
256 return MakeWorkload<ClDepthwiseConvolutionWorkload>(descriptor,
info);
262 return MakeWorkload<ClDequantizeWorkload>(descriptor,
info);
269 return MakeWorkload<NullWorkload, NullWorkload>(descriptor,
info);
275 return MakeWorkload<ClDivisionFloatWorkload, NullWorkload>(descriptor,
info);
289 return std::make_unique<ClAbsWorkload>(absQueueDescriptor,
info);
292 return std::make_unique<ClExpWorkload>(descriptor,
info);
294 return std::make_unique<ClNegWorkload>(descriptor,
info);
298 rsqrtQueueDescriptor.
m_Inputs = descriptor.m_Inputs;
299 rsqrtQueueDescriptor.
m_Outputs = descriptor.m_Outputs;
301 return std::make_unique<ClRsqrtWorkload>(rsqrtQueueDescriptor,
info);
304 return std::make_unique<ClLogicalNotWorkload>(descriptor,
info);
324 return std::make_unique<ClFillWorkload>(descriptor,
info);
330 return MakeWorkload<ClFloorFloatWorkload, NullWorkload>(descriptor,
info);
336 return MakeWorkload<ClFullyConnectedWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
342 return MakeWorkload<ClGatherWorkload>(descriptor,
info);
359 return std::make_unique<CopyMemGenericWorkload>(descriptor,
info);
366 return MakeWorkload<ClInstanceNormalizationWorkload>(descriptor,
info);
372 return MakeWorkload<ClL2NormalizationFloatWorkload, NullWorkload>(descriptor,
info);
381 return std::make_unique<ClLogicalAndWorkload>(descriptor,
info);
383 return std::make_unique<ClLogicalOrWorkload>(descriptor,
info);
395 return std::make_unique<ClLogicalNotWorkload>(descriptor,
info);
404 return MakeWorkload<ClLogSoftmaxWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
410 return MakeWorkload<ClLstmFloatWorkload, NullWorkload>(descriptor,
info);
416 return MakeWorkload<ClMaximumWorkload>(descriptor,
info);
422 return MakeWorkload<ClMeanWorkload>(descriptor,
info);
433 return MakeWorkload<CopyMemGenericWorkload>(descriptor,
info);
444 return std::make_unique<ImportMemGenericWorkload>(descriptor,
info);
456 return MakeWorkload<ClMinimumWorkload>(descriptor,
info);
462 return MakeWorkload<ClMultiplicationWorkload>(descriptor,
info);
468 return MakeWorkload<ClNormalizationFloatWorkload, NullWorkload>(descriptor,
info);
474 return std::make_unique<CopyMemGenericWorkload>(descriptor,
info);
480 return MakeWorkload<ClPadWorkload>(descriptor,
info);
486 return MakeWorkload<ClPermuteWorkload>(descriptor,
info);
492 return MakeWorkload<ClPooling2dWorkload>(descriptor,
info);
498 return MakeWorkload<NullWorkload, NullWorkload>(descriptor,
info);
504 return MakeWorkload<ClPreluWorkload>(descriptor,
info);
510 return std::make_unique<ClQLstmWorkload>(descriptor,
info);
516 return MakeWorkload<ClQuantizeWorkload>(descriptor,
info);
522 return MakeWorkload<ClQuantizedLstmWorkload>(descriptor,
info);
528 return MakeWorkload<ClReshapeWorkload>(descriptor,
info);
534 return MakeWorkload<ClResizeWorkload>(descriptor,
info);
566 return MakeWorkload<ClSliceWorkload>(descriptor,
info);
572 return std::make_unique<ClSoftmaxWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
578 return MakeWorkload<ClSpaceToBatchNdWorkload>(descriptor,
info);
584 return MakeWorkload<ClSpaceToDepthWorkload>(descriptor,
info);
590 return MakeWorkload<ClSplitterWorkload>(descriptor,
info);
596 return MakeWorkload<ClStackWorkload>(descriptor,
info);
602 return MakeWorkload<ClStridedSliceWorkload>(descriptor,
info);
608 return MakeWorkload<ClSubtractionWorkload>(descriptor,
info);
614 return MakeWorkload<ClTransposeWorkload>(descriptor,
info);
621 return MakeWorkload<ClTransposeConvolution2dWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
std::unique_ptr< IWorkload > CreateDetectionPostProcess(const DetectionPostProcessQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< ITensorHandle > CreateSubTensorHandle(ITensorHandle &parent, TensorShape const &subTensorShape, unsigned int const *subTensorOrigin) const override
std::unique_ptr< IWorkload > CreateComparison(const ComparisonQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateGather(const GatherQueueDescriptor &descriptor, const WorkloadInfo &info) const override
ClWorkloadFactory(const std::shared_ptr< ClMemoryManager > &memoryManager)
UnaryOperation m_Operation
Specifies the elementwiseUnary operation to execute.
Interface for a layer that is connectable to other layers via InputSlots and OutputSlots.
std::unique_ptr< IWorkload > CreateDebug(const DebugQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateReshape(const ReshapeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvertFp32ToFp16(const ConvertFp32ToFp16QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDivision(const DivisionQueueDescriptor &descriptor, const WorkloadInfo &info) const override
A ComparisonDescriptor for the ComparisonLayer.
uint32_t m_TargetWidth
Target width value.
DataLayout m_DataLayout
The data layout to be used (NCHW, NHWC).
constexpr const char * ClBackendId()
std::vector< BackendOptions > ModelOptions
std::unique_ptr< IWorkload > CreateStridedSlice(const StridedSliceQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvertFp16ToFp32(const ConvertFp16ToFp32QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateQuantize(const QuantizeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::array< unsigned int, MaxNumOfTensorDimensions > Coordinates
std::unique_ptr< IWorkload > CreateInput(const InputQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateStack(const StackQueueDescriptor &descriptor, const WorkloadInfo &info) const override
ResizeMethod m_Method
The Interpolation method to use (Bilinear, NearestNeighbor).
std::unique_ptr< IWorkload > CreateFullyConnected(const FullyConnectedQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSlice(const SliceQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< ITensorHandle > CreateTensorHandle(const TensorInfo &tensorInfo, const bool IsMemoryManaged=true) const override
std::unique_ptr< IWorkload > CreateQuantizedLstm(const QuantizedLstmQueueDescriptor &descriptor, const WorkloadInfo &info) const override
const BackendId & GetBackendId() const override
Copyright (c) 2020 ARM Limited.
void IgnoreUnused(Ts &&...)
TypedWorkload< QueueDescriptor, armnn::DataType::Float16, armnn::DataType::Float32 > FloatWorkload
std::unique_ptr< IWorkload > CreateAbs(const AbsQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateLstm(const LstmQueueDescriptor &descriptor, const WorkloadInfo &info) const override
LayerDescriptor m_Parameters
LogicalBinaryOperation m_Operation
Specifies the logical operation to execute.
std::unique_ptr< IWorkload > CreateLogicalBinary(const LogicalBinaryQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateLogSoftmax(const LogSoftmaxQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMean(const MeanQueueDescriptor &descriptor, const WorkloadInfo &Info) const override
std::unique_ptr< IWorkload > CreateMinimum(const MinimumQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateTransposeConvolution2d(const TransposeConvolution2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateFill(const FillQueueDescriptor &descriptor, const WorkloadInfo &info) const override
static bool IsLayerSupported(const Layer &layer, Optional< DataType > dataType, std::string &outReasonIfUnsupported)
std::unique_ptr< IWorkload > CreateQLstm(const QLstmQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDepthwiseConvolution2d(const DepthwiseConvolution2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDepthToSpace(const DepthToSpaceQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::shared_ptr< IBackendModelContext > IBackendSpecificModelContextPtr
std::unique_ptr< IWorkload > CreateResize(const ResizeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreatePermute(const PermuteQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMemCopy(const MemCopyQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor &descriptor, const WorkloadInfo &info) const override
uint32_t m_TargetWidth
Target width value.
std::unique_ptr< IWorkload > CreateSoftmax(const SoftmaxQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMaximum(const MaximumQueueDescriptor &descriptor, const WorkloadInfo &info) const override
RuntimeException WrapClError(const cl::Error &clError, const CheckLocation &location)
std::unique_ptr< IWorkload > CreateEqual(const EqualQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor &descriptor, const WorkloadInfo &info) const override
bool IsFastMathEnabled() const
std::unique_ptr< IWorkload > CreateFloor(const FloorQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreatePreCompiled(const PreCompiledQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSubtraction(const SubtractionQueueDescriptor &descriptor, const WorkloadInfo &info) const override
static bool IsLayerSupported(const BackendId &backendId, const IConnectableLayer &layer, Optional< DataType > dataType, std::string &outReasonIfUnsupported)
std::unique_ptr< IWorkload > CreateMultiplication(const MultiplicationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateAddition(const AdditionQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreatePooling2d(const Pooling2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
virtual TensorShape GetShape() const =0
Get the number of elements for each dimension ordered from slowest iterating dimension to fastest ite...
uint32_t m_TargetHeight
Target height value.
uint32_t m_TargetHeight
Target height value.
std::unique_ptr< IWorkload > CreateMerger(const MergerQueueDescriptor &descriptor, const WorkloadInfo &info) const override
TypedWorkload< QueueDescriptor, armnn::DataType::QAsymmU8 > Uint8Workload
std::unique_ptr< IWorkload > CreateRsqrt(const RsqrtQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSplitter(const SplitterQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateGreater(const GreaterQueueDescriptor &descriptor, const WorkloadInfo &info) const override
A ElementwiseUnaryDescriptor for the ElementwiseUnaryLayer.
std::unique_ptr< IWorkload > CreateL2Normalization(const L2NormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateActivation(const ActivationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::vector< ITensorHandle * > m_Outputs
std::unique_ptr< IWorkload > CreatePad(const PadQueueDescriptor &descriptor, const WorkloadInfo &info) const override
unsigned int GetNumDimensions() const
Function that returns the tensor rank.
std::unique_ptr< IWorkload > CreateSpaceToDepth(const SpaceToDepthQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateNormalization(const NormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateInstanceNormalization(const InstanceNormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateTranspose(const TransposeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
Contains information about inputs and outputs to a layer.
std::vector< ITensorHandle * > m_Inputs
std::unique_ptr< IWorkload > CreateBatchNormalization(const BatchNormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConcat(const ConcatQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMemImport(const MemImportQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateLogicalUnary(const ElementwiseUnaryQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateResizeBilinear(const ResizeBilinearQueueDescriptor &descriptor, const WorkloadInfo &info) const override
The ClBackendModelContext is used to pass in CL specific backend ModelOptions.
std::unique_ptr< IWorkload > CreatePrelu(const PreluQueueDescriptor &descriptor, const WorkloadInfo &info) const override
DataLayout m_DataLayout
The data layout to be used (NCHW, NHWC).
std::unique_ptr< IWorkload > CreateDequantize(const DequantizeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConstant(const ConstantQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvolution2d(const Convolution2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateArgMinMax(const ArgMinMaxQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateOutput(const OutputQueueDescriptor &descriptor, const WorkloadInfo &info) const override