ArmNN
 20.11
RefWorkloadFactory.cpp
Go to the documentation of this file.
1 //
2 // Copyright © 2017 Arm Ltd. All rights reserved.
3 // SPDX-License-Identifier: MIT
4 //
5 #include <Layer.hpp>
11 #include "RefWorkloadFactory.hpp"
12 #include "RefBackendId.hpp"
14 #include "RefTensorHandle.hpp"
15 
16 
17 namespace armnn
18 {
19 
20 namespace
21 {
22 static const BackendId s_Id{RefBackendId()};
23 }
24 template <typename F32Workload, typename U8Workload, typename QueueDescriptorType>
25 std::unique_ptr<IWorkload> RefWorkloadFactory::MakeWorkload(const QueueDescriptorType& descriptor,
26  const WorkloadInfo& info) const
27 {
28  return MakeWorkloadHelper<NullWorkload, F32Workload, U8Workload, NullWorkload, NullWorkload, NullWorkload>
29  (descriptor, info);
30 }
31 
32 template <DataType ArmnnType>
34 {
35  auto checkType = [](const TensorInfo& tensorInfo) {return tensorInfo.GetDataType() == ArmnnType;};
36  auto it = std::find_if(std::begin(info.m_InputTensorInfos), std::end(info.m_InputTensorInfos), checkType);
37  if (it != std::end(info.m_InputTensorInfos))
38  {
39  return true;
40  }
41  it = std::find_if(std::begin(info.m_OutputTensorInfos), std::end(info.m_OutputTensorInfos), checkType);
42  if (it != std::end(info.m_OutputTensorInfos))
43  {
44  return true;
45  }
46  return false;
47 }
48 
50 {
51  return IsDataType<DataType::Signed32>(info);
52 }
53 
55 {
56  return IsDataType<DataType::BFloat16>(info);
57 }
58 
60 {
61  return IsDataType<DataType::Float16>(info);
62 }
63 
65 {
66  return IsDataType<DataType::QSymmS16>(info);
67 }
68 
70 {
71  return IsDataType<DataType::QSymmS8>(info);
72 }
73 
75 {
76  return IsDataType<DataType::QAsymmS8>(info);
77 }
78 
80 {
81  return IsDataType<DataType::QAsymmU8>(info);
82 }
83 
84 RefWorkloadFactory::RefWorkloadFactory(const std::shared_ptr<RefMemoryManager>& memoryManager)
85  : m_MemoryManager(memoryManager)
86 {
87 }
88 
90  : m_MemoryManager(new RefMemoryManager())
91 {
92 }
93 
95 {
96  return s_Id;
97 }
98 
100  Optional<DataType> dataType,
101  std::string& outReasonIfUnsupported)
102 {
103  return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported);
104 }
105 
107  Optional<DataType> dataType,
108  std::string& outReasonIfUnsupported,
109  const ModelOptions& modelOptions)
110 {
111  return IWorkloadFactory::IsLayerSupported(s_Id, layer, dataType, outReasonIfUnsupported, modelOptions);
112 }
113 
114 std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
115  const bool isMemoryManaged) const
116 {
117  // For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer
118  // to unmanaged memory. This also ensures memory alignment.
119  IgnoreUnused(isMemoryManaged);
120  return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager);
121 }
122 
123 std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
124  DataLayout dataLayout,
125  const bool isMemoryManaged) const
126 {
127  // For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer
128  // to unmanaged memory. This also ensures memory alignment.
129  IgnoreUnused(isMemoryManaged, dataLayout);
130  return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager);
131 }
132 
133 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateAbs(const AbsQueueDescriptor& descriptor,
134  const WorkloadInfo& info) const
135 {
136  IgnoreUnused(descriptor);
137  ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
138  elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Abs;
139 
140  return CreateElementwiseUnary(elementwiseUnaryDescriptor, info);
141 }
142 
143 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor,
144  const WorkloadInfo& info) const
145 {
146  return std::make_unique<RefActivationWorkload>(descriptor, info);
147 }
148 
149 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor,
150  const WorkloadInfo& info) const
151 {
152  if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
153  {
154  return std::make_unique<RefAdditionWorkload<int32_t>>(descriptor, info);
155  }
156  else
157  {
158  return std::make_unique<RefAdditionWorkload<float>>(descriptor, info);
159  }
160 }
161 
162 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateArgMinMax(const ArgMinMaxQueueDescriptor& descriptor,
163  const WorkloadInfo& info) const
164 {
165  return std::make_unique<RefArgMinMaxWorkload>(descriptor, info);
166 }
167 
169  const BatchNormalizationQueueDescriptor& descriptor,
170  const WorkloadInfo& info) const
171 {
172  return std::make_unique<RefBatchNormalizationWorkload>(descriptor, info);
173 }
174 
175 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor& descriptor,
176  const WorkloadInfo& info) const
177 {
178  return std::make_unique<RefBatchToSpaceNdWorkload>(descriptor, info);
179 }
180 
181 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateComparison(const ComparisonQueueDescriptor& descriptor,
182  const WorkloadInfo& info) const
183 {
184  return std::make_unique<RefComparisonWorkload>(descriptor, info);
185 }
186 
187 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConcat(const ConcatQueueDescriptor& descriptor,
188  const WorkloadInfo& info) const
189 {
190  return std::make_unique<RefConcatWorkload>(descriptor, info);
191 }
192 
193 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor,
194  const WorkloadInfo& info) const
195 {
196  return std::make_unique<RefConstantWorkload>(descriptor, info);
197 }
198 
200  const ConvertBf16ToFp32QueueDescriptor& descriptor,
201  const WorkloadInfo& info) const
202 {
203  return std::make_unique<RefConvertBf16ToFp32Workload>(descriptor, info);
204 }
205 
207  const ConvertFp16ToFp32QueueDescriptor& descriptor,
208  const WorkloadInfo& info) const
209 {
210  return std::make_unique<RefConvertFp16ToFp32Workload>(descriptor, info);
211 }
212 
214  const ConvertFp32ToBf16QueueDescriptor& descriptor,
215  const WorkloadInfo& info) const
216 {
217  return std::make_unique<RefConvertFp32ToBf16Workload>(descriptor, info);
218 }
219 
221  const ConvertFp32ToFp16QueueDescriptor& descriptor,
222  const WorkloadInfo& info) const
223 {
224  return std::make_unique<RefConvertFp32ToFp16Workload>(descriptor, info);
225 }
226 
227 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor,
228  const WorkloadInfo& info) const
229 {
230  return std::make_unique<RefConvolution2dWorkload>(descriptor, info);
231 }
232 
233 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDebug(const DebugQueueDescriptor& descriptor,
234  const WorkloadInfo& info) const
235 {
236  if (IsBFloat16(info))
237  {
238  return std::make_unique<RefDebugBFloat16Workload>(descriptor, info);
239  }
240  if (IsFloat16(info))
241  {
242  return std::make_unique<RefDebugFloat16Workload>(descriptor, info);
243  }
244  if (IsQSymmS16(info))
245  {
246  return std::make_unique<RefDebugQSymmS16Workload>(descriptor, info);
247  }
248  if (IsQSymmS8(info))
249  {
250  return std::make_unique<RefDebugQSymmS8Workload>(descriptor, info);
251  }
252  if (IsQAsymmU8(info))
253  {
254  return std::make_unique<RefDebugQAsymmU8Workload>(descriptor, info);
255  }
256  if (IsQAsymmS8(info))
257  {
258  return std::make_unique<RefDebugQAsymmS8Workload>(descriptor, info);
259  }
260  if (IsSigned32(info))
261  {
262  return std::make_unique<RefDebugSigned32Workload>(descriptor, info);
263  }
264 
265  return MakeWorkload<RefDebugFloat32Workload, RefDebugQAsymmU8Workload>(descriptor, info);
266 }
267 
268 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDepthToSpace(const DepthToSpaceQueueDescriptor& descriptor,
269  const WorkloadInfo& info) const
270 {
271  return std::make_unique<RefDepthToSpaceWorkload>(descriptor, info);
272 }
273 
275  const DepthwiseConvolution2dQueueDescriptor& descriptor,
276  const WorkloadInfo& info) const
277 {
278  return std::make_unique<RefDepthwiseConvolution2dWorkload>(descriptor, info);
279 }
280 
281 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDequantize(const DequantizeQueueDescriptor& descriptor,
282  const WorkloadInfo& info) const
283 {
284  return std::make_unique<RefDequantizeWorkload>(descriptor, info);
285 }
286 
288  const DetectionPostProcessQueueDescriptor& descriptor,
289  const WorkloadInfo& info) const
290 {
291  return std::make_unique<RefDetectionPostProcessWorkload>(descriptor, info);
292 }
293 
294 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateDivision(const DivisionQueueDescriptor& descriptor,
295  const WorkloadInfo& info) const
296 {
297  if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
298  {
299  return std::make_unique<RefDivisionWorkload<int32_t>>(descriptor, info);
300  }
301  else
302  {
303  return std::make_unique<RefDivisionWorkload<float>>(descriptor, info);
304  }
305 }
306 
308  const WorkloadInfo& info) const
309 {
311  {
312  return std::make_unique<RefLogicalUnaryWorkload>(descriptor, info);
313  }
314  return std::make_unique<RefElementwiseUnaryWorkload>(descriptor, info);
315 }
316 
317 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor,
318  const WorkloadInfo& info) const
319 {
320  IgnoreUnused(descriptor);
321  ComparisonQueueDescriptor comparisonDescriptor;
322  comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Equal;
323 
324  return CreateComparison(comparisonDescriptor, info);
325 }
326 
328  const WorkloadInfo& info) const
329 {
330  return MakeWorkload<RefFakeQuantizationFloat32Workload, NullWorkload>(descriptor, info);
331 }
332 
333 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFill(const FillQueueDescriptor& descriptor,
334  const WorkloadInfo& info) const
335 {
336  return std::make_unique<RefFillWorkload>(descriptor, info);
337 }
338 
339 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor,
340  const WorkloadInfo& info) const
341 {
342  if(IsQuantizedType(info.m_InputTensorInfos[0].GetDataType()))
343  {
344  return nullptr;
345  }
346  else
347  {
348  return std::make_unique<RefFloorWorkload>(descriptor, info);
349  }
350 }
351 
352 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateFullyConnected(
353  const FullyConnectedQueueDescriptor& descriptor,
354  const WorkloadInfo& info) const
355 {
356  return std::make_unique<RefFullyConnectedWorkload>(descriptor, info);
357 }
358 
359 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGather(const GatherQueueDescriptor& descriptor,
360  const WorkloadInfo& info) const
361 {
362  return std::make_unique<RefGatherWorkload>(descriptor, info);
363 }
364 
365 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor,
366  const WorkloadInfo& info) const
367 {
368  IgnoreUnused(descriptor);
369  ComparisonQueueDescriptor comparisonDescriptor;
371 
372  return CreateComparison(comparisonDescriptor, info);
373 }
374 
375 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateInput(const InputQueueDescriptor& descriptor,
376  const WorkloadInfo& info) const
377 {
378  if (info.m_InputTensorInfos.empty() )
379  {
380  throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Input cannot be zero length");
381  }
382  if (info.m_OutputTensorInfos.empty())
383  {
384  throw InvalidArgumentException("RefWorkloadFactory::CreateInput: Output cannot be zero length");
385  }
386 
387  if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
388  {
389  throw InvalidArgumentException("RefWorkloadFactory::CreateInput: data input and output differ in byte count.");
390  }
391 
392  return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
393 }
394 
396  const InstanceNormalizationQueueDescriptor& descriptor,
397  const WorkloadInfo& info) const
398 {
399  return std::make_unique<RefInstanceNormalizationWorkload>(descriptor, info);
400 }
401 
403  const WorkloadInfo& info) const
404 {
405  return std::make_unique<RefL2NormalizationWorkload>(descriptor, info);
406 }
407 
408 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLogicalBinary(const LogicalBinaryQueueDescriptor& descriptor,
409  const WorkloadInfo& info) const
410 {
411  return std::make_unique<RefLogicalBinaryWorkload>(descriptor, info);
412 }
413 
414 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLogicalUnary(const ElementwiseUnaryQueueDescriptor& descriptor,
415  const WorkloadInfo& info) const
416 {
417  return std::make_unique<RefLogicalUnaryWorkload>(descriptor, info);
418 }
419 
420 
421 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLogSoftmax(const LogSoftmaxQueueDescriptor& descriptor,
422  const WorkloadInfo& info) const
423 {
424  return std::make_unique<RefLogSoftmaxWorkload>(descriptor, info);
425 }
426 
427 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor,
428  const WorkloadInfo& info) const
429 {
430  return std::make_unique<RefLstmWorkload>(descriptor, info);
431 }
432 
433 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMaximum(const MaximumQueueDescriptor& descriptor,
434  const WorkloadInfo& info) const
435 {
436  if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
437  {
438  return std::make_unique<RefMaximumWorkload<int32_t>>(descriptor, info);
439  }
440  else
441  {
442  return std::make_unique<RefMaximumWorkload<float>>(descriptor, info);
443  }
444 }
445 
446 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMean(const MeanQueueDescriptor& descriptor,
447  const WorkloadInfo& info) const
448 {
449  return std::make_unique<RefMeanWorkload>(descriptor, info);
450 }
451 
452 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor,
453  const WorkloadInfo& info) const
454 {
455  if (descriptor.m_Inputs.empty())
456  {
457  throw InvalidArgumentException("RefWorkloadFactory: CreateMemCopy() expected an input tensor.");
458  }
459  return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
460 }
461 
462 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMemImport(const MemImportQueueDescriptor& descriptor,
463  const WorkloadInfo& info) const
464 {
465  if (descriptor.m_Inputs.empty())
466  {
467  throw InvalidArgumentException("RefWorkloadFactory: CreateMemImport() expected an input tensor.");
468  }
469  return std::make_unique<ImportMemGenericWorkload>(descriptor, info);
470 }
471 
472 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMerger(const MergerQueueDescriptor& descriptor,
473  const WorkloadInfo& info) const
474 {
475  return CreateConcat(descriptor, info);
476 }
477 
478 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMinimum(const MinimumQueueDescriptor& descriptor,
479  const WorkloadInfo& info) const
480 {
481  if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
482  {
483  return std::make_unique<RefMinimumWorkload<int32_t>>(descriptor, info);
484  }
485  else
486  {
487  return std::make_unique<RefMinimumWorkload<float>>(descriptor, info);
488  }
489 }
490 
491 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateMultiplication(const MultiplicationQueueDescriptor& descriptor,
492  const WorkloadInfo& info) const
493 {
494  if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
495  {
496  return std::make_unique<RefMultiplicationWorkload<int32_t>>(descriptor, info);
497  }
498  else
499  {
500  return std::make_unique<RefMultiplicationWorkload<float>>(descriptor, info);
501  }
502 }
503 
504 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& descriptor,
505  const WorkloadInfo& info) const
506 {
507  return std::make_unique<RefNormalizationWorkload>(descriptor, info);
508 }
509 
510 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateOutput(const OutputQueueDescriptor& descriptor,
511  const WorkloadInfo& info) const
512 {
513  if (info.m_InputTensorInfos.empty() )
514  {
515  throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Input cannot be zero length");
516  }
517  if (info.m_OutputTensorInfos.empty())
518  {
519  throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: Output cannot be zero length");
520  }
521  if (info.m_InputTensorInfos[0].GetNumBytes() != info.m_OutputTensorInfos[0].GetNumBytes())
522  {
523  throw InvalidArgumentException("RefWorkloadFactory::CreateOutput: data input and output differ in byte count.");
524  }
525 
526  return std::make_unique<CopyMemGenericWorkload>(descriptor, info);
527 }
528 
529 std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePad(const PadQueueDescriptor& descriptor,
530  const WorkloadInfo& info) const
531 {
532  return std::make_unique<RefPadWorkload>(descriptor, info);
533 }
534 
535 std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePermute(const PermuteQueueDescriptor& descriptor,
536  const WorkloadInfo& info) const
537 {
538  if (IsQSymmS16(info))
539  {
540  return std::make_unique<RefPermuteQSymm16Workload>(descriptor, info);
541  }
542  else if (IsBFloat16(info))
543  {
544  return std::make_unique<RefPermuteBFloat16Workload>(descriptor, info);
545  }
546  else if (IsQAsymmS8(info))
547  {
548  return std::make_unique<RefPermuteQAsymmS8Workload>(descriptor, info);
549  }
551  NullWorkload, NullWorkload, NullWorkload>(descriptor, info);
552 }
553 
554 std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor,
555  const WorkloadInfo& info) const
556 {
557  return std::make_unique<RefPooling2dWorkload>(descriptor, info);
558 }
559 
560 std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePreCompiled(const PreCompiledQueueDescriptor& /*descriptor*/,
561  const WorkloadInfo& /*info*/) const
562 {
563  return nullptr;
564 }
565 
566 std::unique_ptr<IWorkload> RefWorkloadFactory::CreatePrelu(const PreluQueueDescriptor& descriptor,
567  const WorkloadInfo& info) const
568 {
569  return std::make_unique<RefPreluWorkload>(descriptor, info);
570 }
571 
572 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateQLstm(const QLstmQueueDescriptor& descriptor,
573  const WorkloadInfo& info) const
574 {
575  return std::make_unique<RefQLstmWorkload>(descriptor, info);
576 }
577 
578 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateQuantize(const QuantizeQueueDescriptor& descriptor,
579  const WorkloadInfo& info) const
580 {
581  return std::make_unique<RefQuantizeWorkload>(descriptor, info);
582 }
583 
584 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateRank(const RankQueueDescriptor& descriptor,
585  const WorkloadInfo& info) const
586 {
587  return std::make_unique<RefRankWorkload>(descriptor, info);
588 }
589 
590 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor,
591  const WorkloadInfo& info) const
592 {
593  return std::make_unique<RefReshapeWorkload>(descriptor, info);
594 }
595 
596 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResize(const ResizeQueueDescriptor& descriptor,
597  const WorkloadInfo& info) const
598 {
599  return std::make_unique<RefResizeWorkload>(descriptor, info);
600 }
601 
602 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResizeBilinear(const ResizeBilinearQueueDescriptor& descriptor,
603  const WorkloadInfo& info) const
604 {
605  ResizeQueueDescriptor resizeDescriptor;
606  resizeDescriptor.m_Parameters.m_Method = ResizeMethod::Bilinear;
607  resizeDescriptor.m_Parameters.m_DataLayout = descriptor.m_Parameters.m_DataLayout;
608  resizeDescriptor.m_Parameters.m_TargetWidth = descriptor.m_Parameters.m_TargetWidth;
609  resizeDescriptor.m_Parameters.m_TargetHeight = descriptor.m_Parameters.m_TargetHeight;
610 
611  return CreateResize(resizeDescriptor, info);
612 }
613 
614 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& descriptor,
615  const WorkloadInfo& info) const
616 {
617  IgnoreUnused(descriptor);
618  ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
619  elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Rsqrt;
620 
621  return CreateElementwiseUnary(elementwiseUnaryDescriptor, info);
622 }
623 
624 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSlice(const SliceQueueDescriptor& descriptor,
625  const WorkloadInfo& info) const
626 {
627  return std::make_unique<RefSliceWorkload>(descriptor, info);
628 }
629 
630 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor,
631  const WorkloadInfo& info) const
632 {
633  return std::make_unique<RefSoftmaxWorkload>(descriptor, info);
634 }
635 
636 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor& descriptor,
637  const WorkloadInfo& info) const
638 {
639  return std::make_unique<RefSpaceToBatchNdWorkload>(descriptor, info);
640 }
641 
642 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSpaceToDepth(const SpaceToDepthQueueDescriptor& descriptor,
643  const WorkloadInfo& info) const
644 {
645  return std::make_unique<RefSpaceToDepthWorkload>(descriptor, info);
646 }
647 
648 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor,
649  const WorkloadInfo& info) const
650 {
651  return std::make_unique<RefSplitterWorkload>(descriptor, info);
652 }
653 
654 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateStack(const StackQueueDescriptor& descriptor,
655  const WorkloadInfo& info) const
656 {
657  return std::make_unique<RefStackWorkload>(descriptor, info);
658 }
659 
660 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateStridedSlice(const StridedSliceQueueDescriptor& descriptor,
661  const WorkloadInfo& info) const
662 {
663  return std::make_unique<RefStridedSliceWorkload>(descriptor, info);
664 }
665 
666 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateSubtraction(const SubtractionQueueDescriptor& descriptor,
667  const WorkloadInfo& info) const
668 {
669  if (info.m_InputTensorInfos[0].GetDataType() == armnn::DataType::Signed32)
670  {
671  return std::make_unique<RefSubtractionWorkload<int32_t>>(descriptor, info);
672  }
673  else
674  {
675  return std::make_unique<RefSubtractionWorkload<float>>(descriptor, info);
676  }
677 }
678 
679 std::unique_ptr<IWorkload> RefWorkloadFactory::CreateTranspose(const TransposeQueueDescriptor& descriptor,
680  const WorkloadInfo& info) const
681 {
682  if (IsQSymmS16(info))
683  {
684  return std::make_unique<RefTransposeQSymm16Workload>(descriptor, info);
685  }
686  else if (IsBFloat16(info))
687  {
688  return std::make_unique<RefTransposeBFloat16Workload>(descriptor, info);
689  }
690  else if (IsQAsymmS8(info))
691  {
692  return std::make_unique<RefTransposeQAsymmS8Workload>(descriptor, info);
693  }
695  NullWorkload, NullWorkload, NullWorkload>(descriptor, info);
696 }
697 
699  const TransposeConvolution2dQueueDescriptor& descriptor,
700  const WorkloadInfo& info) const
701 {
702  return std::make_unique<RefTransposeConvolution2dWorkload>(descriptor, info);
703 }
704 
705 } // namespace armnn
std::unique_ptr< IWorkload > CreateMemCopy(const MemCopyQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateRsqrt(const RsqrtQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateNormalization(const NormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateArgMinMax(const ArgMinMaxQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateReshape(const ReshapeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateResize(const ResizeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateFakeQuantization(const FakeQuantizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvolution2d(const Convolution2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
UnaryOperation m_Operation
Specifies the elementwiseUnary operation to execute.
std::unique_ptr< IWorkload > CreateConstant(const ConstantQueueDescriptor &descriptor, const WorkloadInfo &info) const override
Interface for a layer that is connectable to other layers via InputSlots and OutputSlots.
Definition: INetwork.hpp:61
std::unique_ptr< IWorkload > CreateEqual(const EqualQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateInput(const InputQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMaximum(const MaximumQueueDescriptor &descriptor, const WorkloadInfo &info) const override
DataLayout
Definition: Types.hpp:50
std::unique_ptr< IWorkload > CreateGather(const GatherQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateQuantize(const QuantizeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
constexpr bool IsQuantizedType()
Definition: TypesUtils.hpp:249
std::unique_ptr< IWorkload > CreateTransposeConvolution2d(const TransposeConvolution2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateStridedSlice(const StridedSliceQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvertFp32ToBf16(const ConvertFp32ToBf16QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSoftmax(const SoftmaxQueueDescriptor &descriptor, const WorkloadInfo &info) const override
uint32_t m_TargetWidth
Target width value.
DataLayout m_DataLayout
The data layout to be used (NCHW, NHWC).
constexpr const char * RefBackendId()
std::vector< BackendOptions > ModelOptions
std::unique_ptr< IWorkload > CreateMultiplication(const MultiplicationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSpaceToDepth(const SpaceToDepthQueueDescriptor &descriptor, const WorkloadInfo &info) const override
ResizeMethod m_Method
The Interpolation method to use (Bilinear, NearestNeighbor).
std::unique_ptr< IWorkload > CreatePad(const PadQueueDescriptor &descriptor, const WorkloadInfo &info) const override
RefPermuteWorkload< DataType::Float16 > RefPermuteFloat16Workload
RefTransposeWorkload< DataType::Float16 > RefTransposeFloat16Workload
std::unique_ptr< IWorkload > CreateMerger(const MergerQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateComparison(const ComparisonQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSpaceToBatchNd(const SpaceToBatchNdQueueDescriptor &descriptor, const WorkloadInfo &info) const override
bool IsQAsymmS8(const WorkloadInfo &info)
std::unique_ptr< IWorkload > CreateLogSoftmax(const LogSoftmaxQueueDescriptor &descriptor, const WorkloadInfo &info) const override
RefPermuteWorkload< DataType::Float32 > RefPermuteFloat32Workload
std::unique_ptr< IWorkload > CreateElementwiseUnary(const ElementwiseUnaryQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConvertFp16ToFp32(const ConvertFp16ToFp32QueueDescriptor &descriptor, const WorkloadInfo &info) const override
Copyright (c) 2020 ARM Limited.
std::unique_ptr< IWorkload > CreateRank(const RankQueueDescriptor &descriptor, const WorkloadInfo &info) const override
void IgnoreUnused(Ts &&...)
bool IsQAsymmU8(const WorkloadInfo &info)
bool IsQSymmS8(const WorkloadInfo &info)
bool IsDataType(const WorkloadInfo &info)
bool IsBFloat16(const WorkloadInfo &info)
std::unique_ptr< IWorkload > CreateConvertFp32ToFp16(const ConvertFp32ToFp16QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateFill(const FillQueueDescriptor &descriptor, const WorkloadInfo &info) const override
const BackendId & GetBackendId() const override
std::unique_ptr< IWorkload > CreateBatchNormalization(const BatchNormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateL2Normalization(const L2NormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateLstm(const LstmQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateConcat(const ConcatQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::vector< TensorInfo > m_InputTensorInfos
bool IsFloat16(const WorkloadInfo &info)
std::unique_ptr< IWorkload > CreateSubtraction(const SubtractionQueueDescriptor &descriptor, const WorkloadInfo &info) const override
RefTransposeWorkload< DataType::Float32 > RefTransposeFloat32Workload
RefTransposeWorkload< DataType::QAsymmU8 > RefTransposeQAsymm8Workload
static bool IsLayerSupported(const Layer &layer, Optional< DataType > dataType, std::string &outReasonIfUnsupported)
std::unique_ptr< IWorkload > CreateDebug(const DebugQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateAddition(const AdditionQueueDescriptor &descriptor, const WorkloadInfo &info) const override
uint32_t m_TargetWidth
Target width value.
bool IsSigned32(const WorkloadInfo &info)
std::vector< TensorInfo > m_OutputTensorInfos
static bool IsLayerSupported(const BackendId &backendId, const IConnectableLayer &layer, Optional< DataType > dataType, std::string &outReasonIfUnsupported)
std::unique_ptr< IWorkload > CreatePooling2d(const Pooling2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
uint32_t m_TargetHeight
Target height value.
uint32_t m_TargetHeight
Target height value.
std::unique_ptr< IWorkload > CreatePrelu(const PreluQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateSplitter(const SplitterQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDequantize(const DequantizeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateFloor(const FloorQueueDescriptor &descriptor, const WorkloadInfo &info) const override
ComparisonOperation m_Operation
Specifies the comparison operation to execute.
Definition: Descriptors.hpp:89
std::unique_ptr< IWorkload > CreateSlice(const SliceQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateFullyConnected(const FullyConnectedQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateBatchToSpaceNd(const BatchToSpaceNdQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateInstanceNormalization(const InstanceNormalizationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMean(const MeanQueueDescriptor &descriptor, const WorkloadInfo &Info) const override
std::unique_ptr< IWorkload > CreateConvertBf16ToFp32(const ConvertBf16ToFp32QueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateStack(const StackQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDivision(const DivisionQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateGreater(const GreaterQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateLogicalUnary(const ElementwiseUnaryQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< ITensorHandle > CreateTensorHandle(const TensorInfo &tensorInfo, const bool IsMemoryManaged=true) const override
std::unique_ptr< IWorkload > CreateLogicalBinary(const LogicalBinaryQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateAbs(const AbsQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateResizeBilinear(const ResizeBilinearQueueDescriptor &descriptor, const WorkloadInfo &info) const override
Contains information about inputs and outputs to a layer.
std::unique_ptr< IWorkload > CreatePermute(const PermuteQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateActivation(const ActivationQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::vector< ITensorHandle * > m_Inputs
RefPermuteWorkload< DataType::QAsymmU8 > RefPermuteQAsymm8Workload
DataLayout m_DataLayout
The data layout to be used (NCHW, NHWC).
std::unique_ptr< IWorkload > CreateMinimum(const MinimumQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateOutput(const OutputQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreatePreCompiled(const PreCompiledQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateTranspose(const TransposeQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDetectionPostProcess(const DetectionPostProcessQueueDescriptor &descriptor, const WorkloadInfo &info) const override
bool IsQSymmS16(const WorkloadInfo &info)
std::unique_ptr< IWorkload > CreateDepthwiseConvolution2d(const DepthwiseConvolution2dQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateQLstm(const QLstmQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateMemImport(const MemImportQueueDescriptor &descriptor, const WorkloadInfo &info) const override
std::unique_ptr< IWorkload > CreateDepthToSpace(const DepthToSpaceQueueDescriptor &descriptor, const WorkloadInfo &info) const override