36 std::string& outReasonIfUnsupported)
43 std::string& outReasonIfUnsupported,
54 void NeonWorkloadFactory::SetNumberOfThreads()
56 if (m_ModelContextPtr)
58 const unsigned int MIN_THREADS = 1;
59 const unsigned int MAX_THREADS = 64;
66 if (numberOfThreads != 0 && numberOfThreads >= MIN_THREADS && numberOfThreads <= MAX_THREADS)
68 arm_compute::Scheduler::get().set_num_threads(numberOfThreads);
74 : m_MemoryManager(memoryManager), m_ModelContextPtr(
IBackendInternal::IBackendSpecificModelContextPtr{})
81 : m_MemoryManager(memoryManager), m_ModelContextPtr(modelContextPtr)
88 unsigned int const* subTensorOrigin)
const 90 const arm_compute::TensorShape shape = armcomputetensorutils::BuildArmComputeTensorShape(subTensorShape);
98 coords.set(i, armnn::numeric_cast<int>(subTensorOrigin[revertedIndex]));
101 const arm_compute::TensorShape parentShape = armcomputetensorutils::BuildArmComputeTensorShape(parent.
GetShape());
102 if (!::arm_compute::error_on_invalid_subtensor(__func__, __FILE__, __LINE__, parentShape, coords, shape))
107 return std::make_unique<NeonSubTensorHandle>(
108 PolymorphicDowncast<IAclTensorHandle*>(&parent), shape, coords);
112 const bool IsMemoryManaged)
const 114 auto tensorHandle = std::make_unique<NeonTensorHandle>(tensorInfo);
117 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
124 const bool IsMemoryManaged)
const 126 auto tensorHandle = std::make_unique<NeonTensorHandle>(tensorInfo, dataLayout);
129 tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
142 auto activationQueueDescriptor = PolymorphicDowncast<const ActivationQueueDescriptor*>(&descriptor);
143 return std::make_unique<NeonActivationWorkload>(*activationQueueDescriptor,
info);
147 auto additionQueueDescriptor = PolymorphicDowncast<const AdditionQueueDescriptor*>(&descriptor);
148 return std::make_unique<NeonAdditionWorkload>(*additionQueueDescriptor,
info);
152 auto argMinMaxQueueDescriptor = PolymorphicDowncast<const ArgMinMaxQueueDescriptor*>(&descriptor);
153 return std::make_unique<NeonArgMinMaxWorkload>(*argMinMaxQueueDescriptor,
info);
157 auto batchNormalizationQueueDescriptor
158 = PolymorphicDowncast<const BatchNormalizationQueueDescriptor*>(&descriptor);
159 return std::make_unique<NeonBatchNormalizationWorkload>(*batchNormalizationQueueDescriptor,
info);
163 auto batchToSpaceNdQueueDescriptor
164 = PolymorphicDowncast<const BatchToSpaceNdQueueDescriptor*>(&descriptor);
165 return std::make_unique<NeonBatchToSpaceNdWorkload>(*batchToSpaceNdQueueDescriptor,
info);
169 auto castQueueDescriptor = PolymorphicDowncast<const CastQueueDescriptor*>(&descriptor);
170 return std::make_unique<NeonCastWorkload>(*castQueueDescriptor,
info);
174 auto channelShuffleQueueDescriptor = PolymorphicDowncast<const ChannelShuffleQueueDescriptor*>(&descriptor);
175 return std::make_unique<NeonChannelShuffleWorkload>(*channelShuffleQueueDescriptor,
info);
179 auto comparisonQueueDescriptor = PolymorphicDowncast<const ComparisonQueueDescriptor*>(&descriptor);
180 return std::make_unique<NeonComparisonWorkload>(*comparisonQueueDescriptor,
info);
184 auto concatQueueDescriptor = PolymorphicDowncast<const ConcatQueueDescriptor*>(&descriptor);
185 return std::make_unique<NeonConcatWorkload>(*concatQueueDescriptor,
info);
189 auto constantQueueDescriptor = PolymorphicDowncast<const ConstantQueueDescriptor*>(&descriptor);
190 return std::make_unique<NeonConstantWorkload>(*constantQueueDescriptor,
info);
194 auto convertBf16ToFp32QueueDescriptor
195 = PolymorphicDowncast<const ConvertBf16ToFp32QueueDescriptor*>(&descriptor);
196 return std::make_unique<NeonConvertBf16ToFp32Workload>(*convertBf16ToFp32QueueDescriptor,
info);
200 auto convertFp16ToFp32QueueDescriptor
201 = PolymorphicDowncast<const ConvertFp16ToFp32QueueDescriptor*>(&descriptor);
202 return std::make_unique<NeonConvertFp16ToFp32Workload>(*convertFp16ToFp32QueueDescriptor,
info);
206 auto convertFp32ToBf16QueueDescriptor
207 = PolymorphicDowncast<const ConvertFp32ToBf16QueueDescriptor*>(&descriptor);
208 return std::make_unique<NeonConvertFp32ToBf16Workload>(*convertFp32ToBf16QueueDescriptor,
info);
212 auto convertFp32ToFp16QueueDescriptor
213 = PolymorphicDowncast<const ConvertFp32ToFp16QueueDescriptor*>(&descriptor);
214 return std::make_unique<NeonConvertFp32ToFp16Workload>(*convertFp32ToFp16QueueDescriptor,
info);
218 auto convolution2dQueueDescriptor = PolymorphicDowncast<const Convolution2dQueueDescriptor*>(&descriptor);
220 bool isFastMathEnabled =
false;
221 if (m_ModelContextPtr)
223 if (m_ModelContextPtr.get() !=
nullptr)
232 return std::make_unique<NeonConvolution2dWorkload>(*convolution2dQueueDescriptor,
234 m_MemoryManager->GetIntraLayerManager(),
239 auto convolution3dQueueDescriptor = PolymorphicDowncast<const Convolution3dQueueDescriptor*>(&descriptor);
241 bool isFastMathEnabled =
false;
242 if (m_ModelContextPtr)
244 if (m_ModelContextPtr.get() !=
nullptr)
253 return std::make_unique<NeonConvolution3dWorkload>(*convolution3dQueueDescriptor,
255 m_MemoryManager->GetIntraLayerManager(),
260 auto debugQueueDescriptor = PolymorphicDowncast<const DebugQueueDescriptor*>(&descriptor);
261 return MakeWorkloadHelper<NullWorkload, NullWorkload>(*debugQueueDescriptor,
info);
265 auto depthToSpaceQueueDescriptor = PolymorphicDowncast<const DepthToSpaceQueueDescriptor*>(&descriptor);
266 return std::make_unique<NeonDepthToSpaceWorkload>(*depthToSpaceQueueDescriptor,
info);
270 auto depthwiseConvolution2dQueueDescriptor
271 = PolymorphicDowncast<const DepthwiseConvolution2dQueueDescriptor*>(&descriptor);
272 return std::make_unique<NeonDepthwiseConvolutionWorkload>(*depthwiseConvolution2dQueueDescriptor,
info);
276 auto dequantizeQueueDescriptor = PolymorphicDowncast<const DequantizeQueueDescriptor*>(&descriptor);
277 return std::make_unique<NeonDequantizeWorkload>(*dequantizeQueueDescriptor,
info);
281 auto detectionPostProcessQueueDescriptor
282 = PolymorphicDowncast<const DetectionPostProcessQueueDescriptor*>(&descriptor);
283 return MakeWorkloadHelper<NullWorkload, NullWorkload>(*detectionPostProcessQueueDescriptor,
info);
287 auto divisionQueueDescriptor = PolymorphicDowncast<const DivisionQueueDescriptor*>(&descriptor);
288 return std::make_unique<NeonDivisionWorkload>(*divisionQueueDescriptor,
info);
292 auto elementwiseUnaryQueueDescriptor
293 = PolymorphicDowncast<const ElementwiseUnaryQueueDescriptor*>(&descriptor);
295 switch(elementwiseUnaryQueueDescriptor->m_Parameters.m_Operation)
300 absQueueDescriptor.
m_Inputs = elementwiseUnaryQueueDescriptor->m_Inputs;
301 absQueueDescriptor.
m_Outputs = elementwiseUnaryQueueDescriptor->m_Outputs;
303 return std::make_unique<NeonAbsWorkload>(absQueueDescriptor,
info);
306 return std::make_unique<NeonExpWorkload>(*elementwiseUnaryQueueDescriptor,
info);
308 return std::make_unique<NeonLogicalNotWorkload>(*elementwiseUnaryQueueDescriptor,
info);
310 return std::make_unique<NeonLogWorkload>(*elementwiseUnaryQueueDescriptor,
info);
312 return std::make_unique<NeonNegWorkload>(*elementwiseUnaryQueueDescriptor,
info);
316 rsqrtQueueDescriptor.
m_Inputs = elementwiseUnaryQueueDescriptor->m_Inputs;
317 rsqrtQueueDescriptor.
m_Outputs = elementwiseUnaryQueueDescriptor->m_Outputs;
319 return std::make_unique<NeonRsqrtWorkload>(rsqrtQueueDescriptor,
info);
322 return std::make_unique<NeonSinWorkload>(*elementwiseUnaryQueueDescriptor,
info);
329 auto fillQueueDescriptor = PolymorphicDowncast<const FillQueueDescriptor*>(&descriptor);
330 return std::make_unique<NeonFillWorkload>(*fillQueueDescriptor,
info);
334 auto floorQueueDescriptor = PolymorphicDowncast<const FloorQueueDescriptor*>(&descriptor);
335 return MakeWorkloadHelper<NeonFloorFloatWorkload, NullWorkload>(*floorQueueDescriptor,
info);
339 auto fullyConnectedQueueDescriptor = PolymorphicDowncast<const FullyConnectedQueueDescriptor*>(&descriptor);
340 return std::make_unique<NeonFullyConnectedWorkload>(*fullyConnectedQueueDescriptor,
342 m_MemoryManager->GetIntraLayerManager());
346 auto gatherQueueDescriptor = PolymorphicDowncast<const GatherQueueDescriptor*>(&descriptor);
347 return std::make_unique<NeonGatherWorkload>(*gatherQueueDescriptor,
info);
351 auto inputQueueDescriptor = PolymorphicDowncast<const InputQueueDescriptor*>(&descriptor);
352 return std::make_unique<CopyMemGenericWorkload>(*inputQueueDescriptor,
info);
356 auto instanceNormalizationQueueDescriptor
357 = PolymorphicDowncast<const InstanceNormalizationQueueDescriptor*>(&descriptor);
358 return std::make_unique<NeonInstanceNormalizationWorkload>(*instanceNormalizationQueueDescriptor,
info);
362 auto l2NormalizationQueueDescriptor
363 = PolymorphicDowncast<const L2NormalizationQueueDescriptor*>(&descriptor);
364 return MakeWorkloadHelper<NeonL2NormalizationFloatWorkload, NullWorkload>
365 (*l2NormalizationQueueDescriptor,
info, m_MemoryManager->GetIntraLayerManager());
369 auto logSoftmaxQueueDescriptor = PolymorphicDowncast<const LogSoftmaxQueueDescriptor*>(&descriptor);
370 return std::make_unique<NeonLogSoftmaxWorkload>(*logSoftmaxQueueDescriptor,
372 m_MemoryManager->GetIntraLayerManager());
376 auto logicalBinaryQueueDescriptor = PolymorphicDowncast<const LogicalBinaryQueueDescriptor*>(&descriptor);
378 switch(logicalBinaryQueueDescriptor->m_Parameters.m_Operation)
381 return std::make_unique<NeonLogicalAndWorkload>(*logicalBinaryQueueDescriptor,
info);
383 return std::make_unique<NeonLogicalOrWorkload>(*logicalBinaryQueueDescriptor,
info);
390 auto lstmQueueDescriptor = PolymorphicDowncast<const LstmQueueDescriptor*>(&descriptor);
391 return MakeWorkloadHelper<NeonLstmFloatWorkload, NullWorkload>(*lstmQueueDescriptor,
info);
395 auto maximumQueueDescriptor = PolymorphicDowncast<const MaximumQueueDescriptor*>(&descriptor);
396 return std::make_unique<NeonMaximumWorkload>(*maximumQueueDescriptor,
info);
400 auto meanQueueDescriptor = PolymorphicDowncast<const MeanQueueDescriptor*>(&descriptor);
401 return std::make_unique<NeonMeanWorkload>(*meanQueueDescriptor,
info);
405 auto memCopyQueueDescriptor = PolymorphicDowncast<const MemCopyQueueDescriptor*>(&descriptor);
406 if (memCopyQueueDescriptor->m_Inputs.empty() || !memCopyQueueDescriptor->m_Inputs[0])
410 return MakeWorkloadHelper<CopyMemGenericWorkload, CopyMemGenericWorkload>(*memCopyQueueDescriptor,
info);
414 auto memImportQueueDescriptor = PolymorphicDowncast<const MemImportQueueDescriptor*>(&descriptor);
415 if (memImportQueueDescriptor->m_Inputs.empty() || !memImportQueueDescriptor->m_Inputs[0])
419 return std::make_unique<ImportMemGenericWorkload>(*memImportQueueDescriptor,
info);
423 auto minimumQueueDescriptor = PolymorphicDowncast<const MinimumQueueDescriptor*>(&descriptor);
424 return std::make_unique<NeonMinimumWorkload>(*minimumQueueDescriptor,
info);
428 auto multiplicationQueueDescriptor = PolymorphicDowncast<const MultiplicationQueueDescriptor*>(&descriptor);
429 return std::make_unique<NeonMultiplicationWorkload>(*multiplicationQueueDescriptor,
info);
433 auto normalizationQueueDescriptor = PolymorphicDowncast<const NormalizationQueueDescriptor*>(&descriptor);
434 return MakeWorkloadHelper<NeonNormalizationFloatWorkload, NullWorkload>
435 (*normalizationQueueDescriptor,
info, m_MemoryManager->GetIntraLayerManager());
439 auto outputQueueDescriptor = PolymorphicDowncast<const OutputQueueDescriptor*>(&descriptor);
440 return std::make_unique<CopyMemGenericWorkload>(*outputQueueDescriptor,
info);
444 auto padQueueDescriptor = PolymorphicDowncast<const PadQueueDescriptor*>(&descriptor);
445 return std::make_unique<NeonPadWorkload>(*padQueueDescriptor,
info);
449 auto permuteQueueDescriptor = PolymorphicDowncast<const PermuteQueueDescriptor*>(&descriptor);
450 return std::make_unique<NeonPermuteWorkload>(*permuteQueueDescriptor,
info);
454 auto pooling2dQueueDescriptor = PolymorphicDowncast<const Pooling2dQueueDescriptor*>(&descriptor);
455 return std::make_unique<NeonPooling2dWorkload>(*pooling2dQueueDescriptor,
info);
459 auto preCompiledQueueDescriptor = PolymorphicDowncast<const PreCompiledQueueDescriptor*>(&descriptor);
460 return MakeWorkloadHelper<NullWorkload, NullWorkload>(*preCompiledQueueDescriptor,
info);
464 auto preluQueueDescriptor = PolymorphicDowncast<const PreluQueueDescriptor*>(&descriptor);
465 return std::make_unique<NeonPreluWorkload>(*preluQueueDescriptor,
info);
469 auto qLstmQueueDescriptor = PolymorphicDowncast<const QLstmQueueDescriptor*>(&descriptor);
470 return std::make_unique<NeonQLstmWorkload>(*qLstmQueueDescriptor,
info);
474 auto quantizeQueueDescriptor = PolymorphicDowncast<const QuantizeQueueDescriptor*>(&descriptor);
475 return std::make_unique<NeonQuantizeWorkload>(*quantizeQueueDescriptor,
info);
479 auto quantizedLstmQueueDescriptor = PolymorphicDowncast<const QuantizedLstmQueueDescriptor*>(&descriptor);
480 return std::make_unique<NeonQuantizedLstmWorkload>(*quantizedLstmQueueDescriptor,
info);
484 auto rankQueueDescriptor = PolymorphicDowncast<const RankQueueDescriptor*>(&descriptor);
485 return std::make_unique<NeonRankWorkload>(*rankQueueDescriptor,
info);
489 auto reduceQueueDescriptor = PolymorphicDowncast<const ReduceQueueDescriptor*>(&descriptor);
490 return std::make_unique<NeonReduceWorkload>(*reduceQueueDescriptor,
info);
494 auto reshapeQueueDescriptor = PolymorphicDowncast<const ReshapeQueueDescriptor*>(&descriptor);
495 return std::make_unique<NeonReshapeWorkload>(*reshapeQueueDescriptor,
info);
499 auto resizeQueueDescriptor = PolymorphicDowncast<const ResizeQueueDescriptor*>(&descriptor);
500 return std::make_unique<NeonResizeWorkload>(*resizeQueueDescriptor,
info);
504 auto sliceQueueDescriptor = PolymorphicDowncast<const SliceQueueDescriptor*>(&descriptor);
505 return std::make_unique<NeonSliceWorkload>(*sliceQueueDescriptor,
info);
509 auto softmaxQueueDescriptor = PolymorphicDowncast<const SoftmaxQueueDescriptor*>(&descriptor);
510 return std::make_unique<NeonSoftmaxWorkload>(*softmaxQueueDescriptor,
512 m_MemoryManager->GetIntraLayerManager());
516 auto spaceToBatchNdQueueDescriptor
517 = PolymorphicDowncast<const SpaceToBatchNdQueueDescriptor*>(&descriptor);
518 return std::make_unique<NeonSpaceToBatchNdWorkload>(*spaceToBatchNdQueueDescriptor,
info);
522 auto spaceToDepthQueueDescriptor = PolymorphicDowncast<const SpaceToDepthQueueDescriptor*>(&descriptor);
523 return std::make_unique<NeonSpaceToDepthWorkload>(*spaceToDepthQueueDescriptor,
info);
527 auto splitterQueueDescriptor = PolymorphicDowncast<const SplitterQueueDescriptor*>(&descriptor);
528 return std::make_unique<NeonSplitterWorkload>(*splitterQueueDescriptor,
info);
532 auto stackQueueDescriptor = PolymorphicDowncast<const StackQueueDescriptor*>(&descriptor);
533 return std::make_unique<NeonStackWorkload>(*stackQueueDescriptor,
info);
537 auto stridedSliceQueueDescriptor = PolymorphicDowncast<const StridedSliceQueueDescriptor*>(&descriptor);
538 return std::make_unique<NeonStridedSliceWorkload>(*stridedSliceQueueDescriptor,
info);
542 auto subtractionQueueDescriptor = PolymorphicDowncast<const SubtractionQueueDescriptor*>(&descriptor);
543 return std::make_unique<NeonSubtractionWorkload>(*subtractionQueueDescriptor,
info);
547 auto transposeQueueDescriptor = PolymorphicDowncast<const TransposeQueueDescriptor*>(&descriptor);
548 return std::make_unique<NeonTransposeWorkload>(*transposeQueueDescriptor,
info);
552 auto transposeConvolution2dQueueDescriptor
553 = PolymorphicDowncast<const TransposeConvolution2dQueueDescriptor*>(&descriptor);
554 return std::make_unique<NeonTransposeConvolution2dWorkload>(*transposeConvolution2dQueueDescriptor,
556 m_MemoryManager->GetIntraLayerManager());
566 return std::make_unique<NeonActivationWorkload>(descriptor,
info);
572 return std::make_unique<NeonAdditionWorkload>(descriptor,
info);
578 return std::make_unique<NeonArgMinMaxWorkload>(descriptor,
info);
584 return std::make_unique<NeonBatchNormalizationWorkload>(descriptor,
info);
590 return std::make_unique<NeonBatchToSpaceNdWorkload>(descriptor,
info);
593 std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateCast(
const CastQueueDescriptor& descriptor,
596 return std::make_unique<NeonCastWorkload>(descriptor,
info);
602 return std::make_unique<NeonChannelShuffleWorkload>(descriptor,
info);
608 return std::make_unique<NeonComparisonWorkload>(descriptor,
info);
614 return std::make_unique<NeonConcatWorkload>(descriptor,
info);
620 return std::make_unique<NeonConstantWorkload>(descriptor,
info);
623 std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateConvertBf16ToFp32(
627 return std::make_unique<NeonConvertBf16ToFp32Workload>(descriptor,
info);
634 return std::make_unique<NeonConvertFp16ToFp32Workload>(descriptor,
info);
637 std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateConvertFp32ToBf16(
641 return std::make_unique<NeonConvertFp32ToBf16Workload>(descriptor,
info);
648 return std::make_unique<NeonConvertFp32ToFp16Workload>(descriptor,
info);
654 bool isFastMathEnabled =
false;
655 if (m_ModelContextPtr)
657 if (m_ModelContextPtr.get() !=
nullptr)
666 return std::make_unique<NeonConvolution2dWorkload>(descriptor,
668 m_MemoryManager->GetIntraLayerManager(),
672 std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateConvolution3d(
675 bool isFastMathEnabled =
false;
676 if (m_ModelContextPtr)
678 if (m_ModelContextPtr.get() !=
nullptr)
687 return std::make_unique<NeonConvolution3dWorkload>(descriptor,
689 m_MemoryManager->GetIntraLayerManager(),
696 return MakeWorkloadHelper<NullWorkload, NullWorkload>(descriptor,
info);
702 return std::make_unique<NeonDepthToSpaceWorkload>(descriptor,
info);
708 return std::make_unique<NeonDepthwiseConvolutionWorkload>(descriptor,
info);
714 return std::make_unique<NeonDequantizeWorkload>(descriptor,
info);
720 return MakeWorkloadHelper<NullWorkload, NullWorkload>(descriptor,
info);
726 return std::make_unique<NeonDivisionWorkload>(descriptor,
info);
740 return std::make_unique<NeonAbsWorkload>(absQueueDescriptor,
info);
743 return std::make_unique<NeonExpWorkload>(descriptor,
info);
745 return std::make_unique<NeonLogicalNotWorkload>(descriptor,
info);
747 return std::make_unique<NeonLogWorkload>(descriptor,
info);
749 return std::make_unique<NeonNegWorkload>(descriptor,
info);
753 rsqrtQueueDescriptor.
m_Inputs = descriptor.m_Inputs;
754 rsqrtQueueDescriptor.
m_Outputs = descriptor.m_Outputs;
756 return std::make_unique<NeonRsqrtWorkload>(rsqrtQueueDescriptor,
info);
759 return std::make_unique<NeonSinWorkload>(descriptor,
info);
765 std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateFill(
const FillQueueDescriptor& descriptor,
768 return std::make_unique<NeonFillWorkload>(descriptor,
info);
774 return MakeWorkloadHelper<NeonFloorFloatWorkload, NullWorkload>(descriptor,
info);
780 return std::make_unique<NeonFullyConnectedWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
786 return std::make_unique<NeonGatherWorkload>(descriptor,
info);
792 return std::make_unique<CopyMemGenericWorkload>(descriptor,
info);
799 return std::make_unique<NeonInstanceNormalizationWorkload>(descriptor,
info);
805 return MakeWorkloadHelper<NeonL2NormalizationFloatWorkload, NullWorkload>(descriptor,
info,
806 m_MemoryManager->GetIntraLayerManager());
812 return std::make_unique<NeonLogSoftmaxWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
821 return std::make_unique<NeonLogicalAndWorkload>(descriptor,
info);
823 return std::make_unique<NeonLogicalOrWorkload>(descriptor,
info);
832 return MakeWorkloadHelper<NeonLstmFloatWorkload, NullWorkload>(descriptor,
info);
838 return std::make_unique<NeonMaximumWorkload>(descriptor,
info);
844 return std::make_unique<NeonMeanWorkload>(descriptor,
info);
855 return MakeWorkloadHelper<CopyMemGenericWorkload, CopyMemGenericWorkload>(descriptor,
info);
866 return std::make_unique<ImportMemGenericWorkload>(descriptor,
info);
872 return std::make_unique<NeonMinimumWorkload>(descriptor,
info);
878 return std::make_unique<NeonMultiplicationWorkload>(descriptor,
info);
884 return MakeWorkloadHelper<NeonNormalizationFloatWorkload, NullWorkload>(descriptor,
info,
885 m_MemoryManager->GetIntraLayerManager());
891 return std::make_unique<CopyMemGenericWorkload>(descriptor,
info);
897 return std::make_unique<NeonPadWorkload>(descriptor,
info);
903 return std::make_unique<NeonPermuteWorkload>(descriptor,
info);
909 return std::make_unique<NeonPooling2dWorkload>(descriptor,
info);
915 return MakeWorkloadHelper<NullWorkload, NullWorkload>(descriptor,
info);
921 return std::make_unique<NeonPreluWorkload>(descriptor,
info);
924 std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateQLstm(
const QLstmQueueDescriptor& descriptor,
927 return std::make_unique<NeonQLstmWorkload>(descriptor,
info);
933 return std::make_unique<NeonQuantizeWorkload>(descriptor,
info);
939 return std::make_unique<NeonQuantizedLstmWorkload>(descriptor,
info);
945 return std::make_unique<NeonRankWorkload>(descriptor,
info);
948 std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateReduce(
const ReduceQueueDescriptor& descriptor,
951 return std::make_unique<NeonReduceWorkload>(descriptor,
info);
957 return std::make_unique<NeonReshapeWorkload>(descriptor,
info);
963 return std::make_unique<NeonResizeWorkload>(descriptor,
info);
969 return std::make_unique<NeonSliceWorkload>(descriptor,
info);
975 return std::make_unique<NeonSoftmaxWorkload>(descriptor,
info, m_MemoryManager->GetIntraLayerManager());
981 return std::make_unique<NeonSpaceToBatchNdWorkload>(descriptor,
info);
987 return std::make_unique<NeonSpaceToDepthWorkload>(descriptor,
info);
993 return std::make_unique<NeonSplitterWorkload>(descriptor,
info);
999 return std::make_unique<NeonStackWorkload>(descriptor,
info);
1005 return std::make_unique<NeonStridedSliceWorkload>(descriptor,
info);
1011 return std::make_unique<NeonSubtractionWorkload>(descriptor,
info);
1017 return std::make_unique<NeonTransposeWorkload>(descriptor,
info);
1024 return std::make_unique<NeonTransposeConvolution2dWorkload>(descriptor,
info,
1025 m_MemoryManager->GetIntraLayerManager());
std::unique_ptr< IWorkload > CreateSubtraction(const SubtractionQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor &descriptor, const WorkloadInfo &info) const override
UnaryOperation m_Operation
Specifies the elementwiseUnary operation to execute.
std::unique_ptr< IWorkload > CreateStridedSlice(const StridedSliceQueueDescriptor &, const WorkloadInfo &) const override
Interface for a layer that is connectable to other layers via InputSlots and OutputSlots.
std::unique_ptr< IWorkload > CreateInput(const InputQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateL2Normalization(const L2NormalizationQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateBatchNormalization(const BatchNormalizationQueueDescriptor &, const WorkloadInfo &) const override
std::vector< BackendOptions > ModelOptions
std::unique_ptr< IWorkload > CreateSoftmax(const SoftmaxQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreatePooling2d(const Pooling2dQueueDescriptor &, const WorkloadInfo &) const override
std::array< unsigned int, MaxNumOfTensorDimensions > Coordinates
constexpr const char * NeonBackendId()
std::unique_ptr< IWorkload > CreateMean(const MeanQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreatePad(const PadQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateLstm(const LstmQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateDepthToSpace(const DepthToSpaceQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateLogSoftmax(const LogSoftmaxQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateComparison(const ComparisonQueueDescriptor &, const WorkloadInfo &) const override
bool IsFastMathEnabled() const
std::unique_ptr< IWorkload > CreateTranspose(const TransposeQueueDescriptor &, const WorkloadInfo &) const override
Copyright (c) 2021 ARM Limited and Contributors.
std::unique_ptr< ITensorHandle > CreateSubTensorHandle(ITensorHandle &parent, TensorShape const &subTensorShape, unsigned int const *subTensorOrigin) const override
unsigned int GetNumberOfThreads() const
std::unique_ptr< IWorkload > CreateDequantize(const DequantizeQueueDescriptor &, const WorkloadInfo &) const override
LayerDescriptor m_Parameters
std::unique_ptr< IWorkload > CreateSplitter(const SplitterQueueDescriptor &, const WorkloadInfo &) const override
LogicalBinaryOperation m_Operation
Specifies the logical operation to execute.
std::unique_ptr< IWorkload > CreateConvertFp32ToFp16(const ConvertFp32ToFp16QueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateNormalization(const NormalizationQueueDescriptor &, const WorkloadInfo &) const override
The NeonBackendModelContext is used to pass in Neon specific backend ModelOptions.
std::unique_ptr< IWorkload > CreateDepthwiseConvolution2d(const DepthwiseConvolution2dQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateInstanceNormalization(const InstanceNormalizationQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateTransposeConvolution2d(const TransposeConvolution2dQueueDescriptor &, const WorkloadInfo &) const override
NeonWorkloadFactory(const std::shared_ptr< NeonMemoryManager > &memoryManager)
static bool IsLayerSupported(const Layer &layer, Optional< DataType > dataType, std::string &outReasonIfUnsupported)
std::unique_ptr< IWorkload > CreateSpaceToDepth(const SpaceToDepthQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateConcat(const ConcatQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreatePrelu(const PreluQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateArgMinMax(const ArgMinMaxQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreatePermute(const PermuteQueueDescriptor &, const WorkloadInfo &) const override
std::shared_ptr< IBackendModelContext > IBackendSpecificModelContextPtr
std::unique_ptr< IWorkload > CreateDebug(const DebugQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateConstant(const ConstantQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateConvolution2d(const Convolution2dQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateFullyConnected(const FullyConnectedQueueDescriptor &, const WorkloadInfo &) const override
const BackendId & GetBackendId() const override
std::unique_ptr< IWorkload > CreateMinimum(const MinimumQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateReshape(const ReshapeQueueDescriptor &, const WorkloadInfo &) const override
static bool IsLayerSupported(const BackendId &backendId, const IConnectableLayer &layer, Optional< DataType > dataType, std::string &outReasonIfUnsupported)
std::unique_ptr< IWorkload > CreateFloor(const FloorQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateRank(const RankQueueDescriptor &, const WorkloadInfo &) const override
virtual TensorShape GetShape() const =0
Get the number of elements for each dimension ordered from slowest iterating dimension to fastest ite...
std::unique_ptr< IWorkload > CreateDetectionPostProcess(const DetectionPostProcessQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateQuantize(const QuantizeQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateAddition(const AdditionQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateGather(const GatherQueueDescriptor &, const WorkloadInfo &) const override
std::vector< ITensorHandle * > m_Outputs
std::unique_ptr< IWorkload > CreateDivision(const DivisionQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreatePreCompiled(const PreCompiledQueueDescriptor &, const WorkloadInfo &) const override
unsigned int GetNumDimensions() const
Function that returns the tensor rank.
std::unique_ptr< IWorkload > CreateWorkload(LayerType type, const QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateStack(const StackQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateOutput(const OutputQueueDescriptor &, const WorkloadInfo &) const override
Contains information about TensorInfos of a layer.
std::unique_ptr< IWorkload > CreateResize(const ResizeQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateSlice(const SliceQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateConvertFp16ToFp32(const ConvertFp16ToFp32QueueDescriptor &, const WorkloadInfo &) const override
std::vector< ITensorHandle * > m_Inputs
std::unique_ptr< IWorkload > CreateMultiplication(const MultiplicationQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateMaximum(const MaximumQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateActivation(const ActivationQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateMemImport(const MemImportQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< IWorkload > CreateMemCopy(const MemCopyQueueDescriptor &, const WorkloadInfo &) const override
Depthwise Convolution 2D layer workload data.
std::unique_ptr< IWorkload > CreateQuantizedLstm(const QuantizedLstmQueueDescriptor &, const WorkloadInfo &) const override
std::unique_ptr< ITensorHandle > CreateTensorHandle(const TensorInfo &tensorInfo, const bool IsMemoryManaged=true) const override
LayerType
When adding a new layer, adapt also the LastLayer enum value in the enum class LayerType below...