36 std::string& outReasonIfUnsupported)
43 std::string& outReasonIfUnsupported,
55 : m_MemoryManager(memoryManager), m_ModelContextPtr(
IBackendInternal::IBackendSpecificModelContextPtr{})
61 : m_MemoryManager(memoryManager), m_ModelContextPtr(modelContextPtr)
67 unsigned int const* subTensorOrigin)
const 69 const arm_compute::TensorShape shape = armcomputetensorutils::BuildArmComputeTensorShape(subTensorShape);
77 coords.set(i, armnn::numeric_cast<int>(subTensorOrigin[revertedIndex]));
80 const arm_compute::TensorShape parentShape = armcomputetensorutils::BuildArmComputeTensorShape(parent.
GetShape());
81 if (!::arm_compute::error_on_invalid_subtensor(__func__, __FILE__, __LINE__, parentShape, coords, shape))
86 return std::make_unique<NeonSubTensorHandle>(
87 PolymorphicDowncast<IAclTensorHandle*>(&parent), shape, coords);
91 const bool IsMemoryManaged)
const 93 auto tensorHandle = std::make_unique<NeonTensorHandle>(tensorInfo);
96 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
103 const bool IsMemoryManaged)
const 105 auto tensorHandle = std::make_unique<NeonTensorHandle>(tensorInfo, dataLayout);
108 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
127 return std::make_unique<NeonActivationWorkload>(descriptor,
info);
133 return std::make_unique<NeonAdditionWorkload>(descriptor,
info);
139 return std::make_unique<NeonArgMinMaxWorkload>(descriptor,
info);
145 return std::make_unique<NeonBatchNormalizationWorkload>(descriptor,
info);
151 return std::make_unique<NeonBatchToSpaceNdWorkload>(descriptor,
info);
157 return std::make_unique<NeonComparisonWorkload>(descriptor,
info);
163 return std::make_unique<NeonConcatWorkload>(descriptor,
info);
169 return std::make_unique<NeonConstantWorkload>(descriptor,
info);
176 return std::make_unique<NeonConvertBf16ToFp32Workload>(descriptor,
info);
183 return std::make_unique<NeonConvertFp16ToFp32Workload>(descriptor,
info);
190 return std::make_unique<NeonConvertFp32ToBf16Workload>(descriptor,
info);
197 return std::make_unique<NeonConvertFp32ToFp16Workload>(descriptor,
info);
203 bool isFastMathEnabled =
false;
204 if (m_ModelContextPtr)
206 if (m_ModelContextPtr.get() !=
nullptr)
215 return std::make_unique<NeonConvolution2dWorkload>(descriptor,
217 m_MemoryManager->GetIntraLayerManager(),
224 return MakeWorkloadHelper<NullWorkload, NullWorkload>(descriptor,
info);
230 return std::make_unique<NeonDepthToSpaceWorkload>(descriptor,
info);
236 return std::make_unique<NeonDepthwiseConvolutionWorkload>(descriptor,
info);
242 return std::make_unique<NeonDequantizeWorkload>(descriptor,
info);
248 return MakeWorkloadHelper<NullWorkload, NullWorkload>(descriptor,
info);
254 return std::make_unique<NeonDivisionWorkload>(descriptor,
info);
268 return std::make_unique<NeonAbsWorkload>(absQueueDescriptor,
info);
276 return std::make_unique<NeonRsqrtWorkload>(rsqrtQueueDescriptor,
info);
279 return std::make_unique<NeonNegWorkload>(descriptor,
info);
281 return std::make_unique<NeonExpWorkload>(descriptor,
info);
283 return std::make_unique<NeonLogicalNotWorkload>(descriptor,
info);
303 return std::make_unique<NeonFillWorkload>(descriptor,
info);
309 return MakeWorkloadHelper<NeonFloorFloatWorkload, NullWorkload>(descriptor,
info);
315 return std::make_unique<NeonFullyConnectedWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
321 return std::make_unique<NeonGatherWorkload>(descriptor,
info);
338 return std::make_unique<CopyMemGenericWorkload>(descriptor,
info);
345 return std::make_unique<NeonInstanceNormalizationWorkload>(descriptor,
info);
351 return MakeWorkloadHelper<NeonL2NormalizationFloatWorkload, NullWorkload>(descriptor,
info,
352 m_MemoryManager->GetIntraLayerManager());
358 return std::make_unique<NeonLogSoftmaxWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
367 return std::make_unique<NeonLogicalAndWorkload>(descriptor,
info);
369 return std::make_unique<NeonLogicalOrWorkload>(descriptor,
info);
381 return std::make_unique<NeonLogicalNotWorkload>(descriptor,
info);
390 return MakeWorkloadHelper<NeonLstmFloatWorkload, NullWorkload>(descriptor,
info);
396 return std::make_unique<NeonMaximumWorkload>(descriptor,
info);
402 return std::make_unique<NeonMeanWorkload>(descriptor,
info);
413 return MakeWorkloadHelper<CopyMemGenericWorkload, CopyMemGenericWorkload>(descriptor,
info);
424 return std::make_unique<ImportMemGenericWorkload>(descriptor,
info);
436 return std::make_unique<NeonMinimumWorkload>(descriptor,
info);
442 return std::make_unique<NeonMultiplicationWorkload>(descriptor,
info);
448 return MakeWorkloadHelper<NeonNormalizationFloatWorkload, NullWorkload>(descriptor,
info,
449 m_MemoryManager->GetIntraLayerManager());
455 return std::make_unique<CopyMemGenericWorkload>(descriptor,
info);
461 return std::make_unique<NeonPadWorkload>(descriptor,
info);
467 return std::make_unique<NeonPermuteWorkload>(descriptor,
info);
473 return std::make_unique<NeonPooling2dWorkload>(descriptor,
info);
479 return MakeWorkloadHelper<NullWorkload, NullWorkload>(descriptor,
info);
485 return std::make_unique<NeonPreluWorkload>(descriptor,
info);
491 return std::make_unique<NeonQLstmWorkload>(descriptor,
info);
497 return std::make_unique<NeonQuantizeWorkload>(descriptor,
info);
503 return std::make_unique<NeonQuantizedLstmWorkload>(descriptor,
info);
509 return std::make_unique<NeonReshapeWorkload>(descriptor,
info);
515 return std::make_unique<NeonResizeWorkload>(descriptor,
info);
547 return std::make_unique<NeonSliceWorkload>(descriptor,
info);
553 return std::make_unique<NeonSoftmaxWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
559 return std::make_unique<NeonSpaceToBatchNdWorkload>(descriptor,
info);
565 return std::make_unique<NeonSpaceToDepthWorkload>(descriptor,
info);
571 return std::make_unique<NeonSplitterWorkload>(descriptor,
info);
577 return std::make_unique<NeonStackWorkload>(descriptor,
info);
583 return std::make_unique<NeonStridedSliceWorkload>(descriptor,
info);
589 return std::make_unique<NeonSubtractionWorkload>(descriptor,
info);
595 return std::make_unique<NeonTransposeWorkload>(descriptor,
info);
602 return std::make_unique<NeonTransposeConvolution2dWorkload>(descriptor,
info,
603 m_MemoryManager->GetIntraLayerManager());
std::unique_ptr< IWorkload > CreateDepthToSpace(const DepthToSpaceQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDebug(const DebugQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateLstm(const LstmQueueDescriptor &descriptor, const WorkloadInfo &info) const override
UnaryOperation m_Operation
Specifies the elementwiseUnary operation to execute.
Interface for a layer that is connectable to other layers via InputSlots and OutputSlots.
std::unique_ptr< IWorkload > CreateLogSoftmax(const LogSoftmaxQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateGather(const GatherQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateNormalization(const NormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateFullyConnected(const FullyConnectedQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateComparison(const ComparisonQueueDescriptor &descriptor, const WorkloadInfo &Info) const override
std::unique_ptr< IWorkload > CreateRsqrt(const RsqrtQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreatePad(const PadQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateTransposeConvolution2d(const TransposeConvolution2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
A ComparisonDescriptor for the ComparisonLayer.
uint32_t m_TargetWidth
Target width value.
DataLayout m_DataLayout
The data layout to be used (NCHW, NHWC).
std::vector< BackendOptions > ModelOptions
std::unique_ptr< IWorkload > CreateGreater(const GreaterQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMemImport(const MemImportQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::array< unsigned int, MaxNumOfTensorDimensions > Coordinates
constexpr const char * NeonBackendId()
std::unique_ptr< IWorkload > CreateL2Normalization(const L2NormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvertFp16ToFp32(const ConvertFp16ToFp32QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateStridedSlice(const StridedSliceQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMinimum(const MinimumQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMultiplication(const MultiplicationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateQuantize(const QuantizeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
bool IsFastMathEnabled() const
Copyright (c) 2020 ARM Limited.
void IgnoreUnused(Ts &&...)
std::unique_ptr< ITensorHandle > CreateSubTensorHandle(ITensorHandle &parent, TensorShape const &subTensorShape, unsigned int const *subTensorOrigin) const override
std::unique_ptr< IWorkload > CreateLogicalUnary(const ElementwiseUnaryQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvolution2d(const Convolution2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
LayerDescriptor m_Parameters
LogicalBinaryOperation m_Operation
Specifies the logical operation to execute.
The NeonBackendModelContext is used to pass in Neon specific backend ModelOptions.
NeonWorkloadFactory(const std::shared_ptr< NeonMemoryManager > &memoryManager)
std::unique_ptr< IWorkload > CreateReshape(const ReshapeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
static bool IsLayerSupported(const Layer &layer, Optional< DataType > dataType, std::string &outReasonIfUnsupported)
std::unique_ptr< IWorkload > CreateAbs(const AbsQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvertFp32ToFp16(const ConvertFp32ToFp16QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSoftmax(const SoftmaxQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateResizeBilinear(const ResizeBilinearQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateEqual(const EqualQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSlice(const SliceQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::shared_ptr< IBackendModelContext > IBackendSpecificModelContextPtr
std::unique_ptr< IWorkload > CreateQuantizedLstm(const QuantizedLstmQueueDescriptor &descriptor, const WorkloadInfo &info) const override
uint32_t m_TargetWidth
Target width value.
const BackendId & GetBackendId() const override
std::unique_ptr< IWorkload > CreateSplitter(const SplitterQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateFill(const FillQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDequantize(const DequantizeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMemCopy(const MemCopyQueueDescriptor &descriptor, const WorkloadInfo &info) const override
static bool IsLayerSupported(const BackendId &backendId, const IConnectableLayer &layer, Optional< DataType > dataType, std::string &outReasonIfUnsupported)
virtual TensorShape GetShape() const =0
Get the number of elements for each dimension ordered from slowest iterating dimension to fastest ite...
std::unique_ptr< IWorkload > CreateInstanceNormalization(const InstanceNormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
uint32_t m_TargetHeight
Target height value.
uint32_t m_TargetHeight
Target height value.
std::unique_ptr< IWorkload > CreatePreCompiled(const PreCompiledQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateLogicalBinary(const LogicalBinaryQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateOutput(const OutputQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvertFp32ToBf16(const ConvertFp32ToBf16QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConstant(const ConstantQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateStack(const StackQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateFloor(const FloorQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateActivation(const ActivationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDepthwiseConvolution2d(const DepthwiseConvolution2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
A ElementwiseUnaryDescriptor for the ElementwiseUnaryLayer.
std::vector< ITensorHandle * > m_Outputs
std::unique_ptr< IWorkload > CreateMerger(const MergerQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDivision(const DivisionQueueDescriptor &descriptor, const WorkloadInfo &info) const override
unsigned int GetNumDimensions() const
Function that returns the tensor rank.
std::unique_ptr< IWorkload > CreateInput(const InputQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateQLstm(const QLstmQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateBatchNormalization(const BatchNormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor &descriptor, const WorkloadInfo &Info) const override
Contains information about inputs and outputs to a layer.
std::unique_ptr< IWorkload > CreateResize(const ResizeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::vector< ITensorHandle * > m_Inputs
std::unique_ptr< IWorkload > CreateConcat(const ConcatQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreatePooling2d(const Pooling2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor &descriptor, const WorkloadInfo &Info) const override
std::unique_ptr< IWorkload > CreateAddition(const AdditionQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDetectionPostProcess(const DetectionPostProcessQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSubtraction(const SubtractionQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMaximum(const MaximumQueueDescriptor &descriptor, const WorkloadInfo &info) const override
DataLayout m_DataLayout
The data layout to be used (NCHW, NHWC).
std::unique_ptr< IWorkload > CreateMean(const MeanQueueDescriptor &descriptor, const WorkloadInfo &Info) const override
std::unique_ptr< IWorkload > CreateConvertBf16ToFp32(const ConvertBf16ToFp32QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateArgMinMax(const ArgMinMaxQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreatePrelu(const PreluQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateTranspose(const TransposeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSpaceToDepth(const SpaceToDepthQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreatePermute(const PermuteQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< ITensorHandle > CreateTensorHandle(const TensorInfo &tensorInfo, const bool IsMemoryManaged=true) const override