ArmNN
 22.05
NeonLayerSupport.cpp
Go to the documentation of this file.
1 //
2 // Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
3 // SPDX-License-Identifier: MIT
4 //
5 
6 #include "NeonLayerSupport.hpp"
7 #include "NeonBackendId.hpp"
9 
10 #include <armnn/Exceptions.hpp>
11 #include <armnn/Tensor.hpp>
12 #include <armnn/Types.hpp>
14 
15 #include <InternalTypes.hpp>
16 #include <LayerSupportCommon.hpp>
19 
20 #if defined(ARMCOMPUTENEON_ENABLED)
84 #endif
85 
86 namespace armnn
87 {
88 
89 namespace
90 {
91 
92 template< typename ... Args>
93 bool IsNeonBackendSupported(Optional<std::string&> reasonIfUnsupported, Args... args)
94 {
95  IgnoreUnused(reasonIfUnsupported, (args)...);
96 #if defined(ARMCOMPUTENEON_ENABLED)
97  return true;
98 #else
99  SetValueChecked(reasonIfUnsupported, "The armnn library has been built without NEON support");
100  return false;
101 #endif
102 }
103 
104 template<typename FloatFunc, typename Uint8Func, typename ... Params>
105 bool IsSupportedForDataTypeNeon(Optional<std::string&> reasonIfUnsupported,
106  DataType dataType,
107  FloatFunc floatFuncPtr,
108  Uint8Func uint8FuncPtr,
109  Params&&... params)
110 {
111  return IsNeonBackendSupported(reasonIfUnsupported) &&
112  IsSupportedForDataTypeGeneric(reasonIfUnsupported,
113  dataType,
114  floatFuncPtr,
115  floatFuncPtr,
116  uint8FuncPtr,
117  &FalseFunc<>,
118  &FalseFunc<>,
119  std::forward<Params>(params)...);
120 }
121 
122 #if defined(ARMCOMPUTENEON_ENABLED)
123 template<class FuncType, class... Args>
124 inline bool IsWorkloadSupported(FuncType& func, Optional<std::string&> reasonIfUnsupported, Args&&... args)
125 {
126  arm_compute::Status aclStatus = func(std::forward<Args>(args)...);
127  const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
128  if (!supported && reasonIfUnsupported)
129  {
130  reasonIfUnsupported.value() = aclStatus.error_description();
131  }
132  return supported;
133 }
134 
135 #define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
136  return IsWorkloadSupported(func, reasonIfUnsupported, __VA_ARGS__);
137 #else
138 #define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
139  return IsNeonBackendSupported(reasonIfUnsupported, __VA_ARGS__);
140 #endif
141 } // anonymous namespace
142 
144  : m_ModelContextPtr(modelContextPtr)
145 {
146 }
147 
149  : m_ModelContextPtr(nullptr)
150 {
151 }
152 
154  const std::vector<TensorInfo>& infos,
155  const BaseDescriptor& descriptor,
156  const Optional<LstmInputParamsInfo>& lstmParamsInfo,
157  const Optional<QuantizedLstmInputParamsInfo>& quantizedLstmParamsInfo,
158  Optional<std::string&> reasonIfUnsupported) const
159 {
160  switch (type)
161  {
163  return IsActivationSupported(infos[0],
164  infos[1],
165  *(PolymorphicDowncast<const ActivationDescriptor*>(&descriptor)),
166  reasonIfUnsupported);
167  case LayerType::Addition:
168  return IsAdditionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
170  return IsArgMinMaxSupported(infos[0],
171  infos[1],
172  *(PolymorphicDowncast<const ArgMinMaxDescriptor*>(&descriptor)),
173  reasonIfUnsupported);
175  return IsBatchNormalizationSupported(infos[0],
176  infos[1],
177  infos[2],
178  infos[3],
179  infos[4],
180  infos[5],
181  *(PolymorphicDowncast<const BatchNormalizationDescriptor*>
182  (&descriptor)),
183  reasonIfUnsupported);
185  return IsBatchToSpaceNdSupported(infos[0],
186  infos[1],
187  *(PolymorphicDowncast<const BatchToSpaceNdDescriptor*>(&descriptor)),
188  reasonIfUnsupported);
189  case LayerType::Cast:
190  return IsCastSupported(infos[0], infos[1], reasonIfUnsupported);
192  return IsChannelShuffleSupported(infos[0],
193  infos[1],
194  *(PolymorphicDowncast<const ChannelShuffleDescriptor*>(&descriptor)),
195  reasonIfUnsupported);
197  return IsComparisonSupported(infos[0],
198  infos[1],
199  infos[2],
200  *(PolymorphicDowncast<const ComparisonDescriptor*>(&descriptor)),
201  reasonIfUnsupported);
202  case LayerType::Concat:
203  {
204  std::vector<const TensorInfo*> inputInfos;
205  for (uint32_t i = 0; i < (infos.size() - 1); i++)
206  {
207  inputInfos.push_back(&infos[i]);
208  }
209  return IsConcatSupported(inputInfos,
210  infos[infos.size() - 1],
211  *(PolymorphicDowncast<const OriginsDescriptor*>(&descriptor)),
213  }
214  case LayerType::Constant:
215  return IsConstantSupported(infos[0], reasonIfUnsupported);
217  return IsConvertBf16ToFp32Supported(infos[0], infos[1], reasonIfUnsupported);
219  return IsConvertFp16ToFp32Supported(infos[0], infos[1], reasonIfUnsupported);
221  return IsConvertFp32ToBf16Supported(infos[0], infos[1], reasonIfUnsupported);
223  return IsConvertFp32ToFp16Supported(infos[0], infos[1], reasonIfUnsupported);
225  {
226  if (infos.size() != 4)
227  {
228  throw InvalidArgumentException("Invalid number of TransposeConvolution2d TensorInfos. "
229  "TensorInfos should be of format: {input, output, weights, biases}.");
230  }
231 
232  auto desc = *(PolymorphicDowncast<const Convolution2dDescriptor*>(&descriptor));
233  if (infos[3] == TensorInfo())
234  {
235  return IsConvolution2dSupported(infos[0],
236  infos[1],
237  desc,
238  infos[2],
239  EmptyOptional(),
240  reasonIfUnsupported);
241  }
242  else
243  {
244  return IsConvolution2dSupported(infos[0],
245  infos[1],
246  desc,
247  infos[2],
248  infos[3],
249  reasonIfUnsupported);
250  }
251  }
253  {
254  if (infos.size() != 4)
255  {
256  throw InvalidArgumentException("Invalid number of Convolution3d TensorInfos. "
257  "TensorInfos should be of format: {input, output, weights, biases}.");
258  }
259 
260  auto desc = *(PolymorphicDowncast<const Convolution3dDescriptor*>(&descriptor));
261  if (infos[3] == TensorInfo())
262  {
263  return IsConvolution3dSupported(infos[0],
264  infos[1],
265  desc,
266  infos[2],
267  EmptyOptional(),
268  reasonIfUnsupported);
269  }
270  else
271  {
272  return IsConvolution3dSupported(infos[0],
273  infos[1],
274  desc,
275  infos[2],
276  infos[3],
277  reasonIfUnsupported);
278  }
279  }
281  return IsDepthToSpaceSupported(infos[0],
282  infos[1],
283  *(PolymorphicDowncast<const DepthToSpaceDescriptor*>(&descriptor)),
284  reasonIfUnsupported);
286  {
287  if (infos.size() != 4)
288  {
289  throw InvalidArgumentException("Invalid number of DepthwiseConvolution2d TensorInfos. "
290  "TensorInfos should be of format: {input, output, weights, biases}.");
291  }
292 
293  auto desc = *(PolymorphicDowncast<const DepthwiseConvolution2dDescriptor*>(&descriptor));
294  if (infos[3] == TensorInfo())
295  {
296  return IsDepthwiseConvolutionSupported(infos[0],
297  infos[1],
298  desc,
299  infos[2],
300  EmptyOptional(),
301  reasonIfUnsupported);
302  }
303  else
304  {
305  return IsDepthwiseConvolutionSupported(infos[0],
306  infos[1],
307  desc,
308  infos[2],
309  infos[3],
310  reasonIfUnsupported);
311  }
312  }
314  return IsDequantizeSupported(infos[0], infos[1], reasonIfUnsupported);
316  {
317  auto desc = *(PolymorphicDowncast<const DetectionPostProcessDescriptor*>(&descriptor));
319  infos[1],
320  infos[2],
321  infos[3],
322  infos[4],
323  infos[5],
324  infos[6],
325  desc,
326  reasonIfUnsupported);
327  }
328  case LayerType::Division:
329  return IsDivisionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
331  return IsElementwiseUnarySupported(infos[0],
332  infos[1],
333  *(PolymorphicDowncast<const ElementwiseUnaryDescriptor*>(&descriptor)),
334  reasonIfUnsupported);
335  case LayerType::Fill:
336  return IsFillSupported(infos[0],
337  infos[1],
338  *(PolymorphicDowncast<const FillDescriptor*>(&descriptor)),
339  reasonIfUnsupported);
340  case LayerType::Floor:
341  return IsFloorSupported(infos[0], infos[1], reasonIfUnsupported);
343  return IsFullyConnectedSupported(infos[0],
344  infos[1],
345  infos[2],
346  infos[3],
347  *(PolymorphicDowncast<const FullyConnectedDescriptor*>(&descriptor)),
348  reasonIfUnsupported);
349  case LayerType::Gather:
350  return IsGatherSupported(infos[0],
351  infos[1],
352  infos[2],
353  *(PolymorphicDowncast<const GatherDescriptor*>(&descriptor)),
354  reasonIfUnsupported);
355  case LayerType::GatherNd:
356  return IsGatherNdSupported(infos[0],
357  infos[1],
358  infos[2],
359  reasonIfUnsupported);
360  case LayerType::Input:
361  return IsInputSupported(infos[0], reasonIfUnsupported);
363  return IsInstanceNormalizationSupported(infos[0],
364  infos[1],
365  *(PolymorphicDowncast<const InstanceNormalizationDescriptor*>
366  (&descriptor)),
367  reasonIfUnsupported);
369  return IsL2NormalizationSupported(infos[0],
370  infos[1],
371  *(PolymorphicDowncast<const L2NormalizationDescriptor*>(&descriptor)),
372  reasonIfUnsupported);
374  return IsLogicalBinarySupported(infos[0],
375  infos[1],
376  infos[2],
377  *(PolymorphicDowncast<const LogicalBinaryDescriptor*>(&descriptor)),
378  reasonIfUnsupported);
380  return IsLogSoftmaxSupported(infos[0],
381  infos[1],
382  *(PolymorphicDowncast<const LogSoftmaxDescriptor*>(&descriptor)),
383  reasonIfUnsupported);
384  case LayerType::Lstm:
385  return IsLstmSupported(infos[0],
386  infos[1],
387  infos[2],
388  infos[3],
389  infos[4],
390  infos[5],
391  infos[6],
392  *(PolymorphicDowncast<const LstmDescriptor*>(&descriptor)),
393  lstmParamsInfo.value(),
395  case LayerType::Map:
396  return true;
397  case LayerType::Maximum:
398  return IsMaximumSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
399  case LayerType::Mean:
400  return IsMeanSupported(infos[0],
401  infos[1],
402  *(PolymorphicDowncast<const MeanDescriptor*>(&descriptor)),
403  reasonIfUnsupported);
404  case LayerType::MemCopy:
405  return LayerSupportBase::IsMemCopySupported(infos[0], infos[1], reasonIfUnsupported);
407  return LayerSupportBase::IsMemImportSupported(infos[0], infos[1], reasonIfUnsupported);
408  case LayerType::Merge:
409  return LayerSupportBase::IsMergeSupported(infos[0],
410  infos[1],
411  infos[2],
412  reasonIfUnsupported);
413  case LayerType::Minimum:
414  return IsMinimumSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
416  return IsMultiplicationSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
418  return IsNormalizationSupported(infos[0],
419  infos[1],
420  *(PolymorphicDowncast<const NormalizationDescriptor*>(&descriptor)),
421  reasonIfUnsupported);
422  case LayerType::Output:
423  return IsOutputSupported(infos[0], reasonIfUnsupported);
424  case LayerType::Pad:
425  return IsPadSupported(infos[0],
426  infos[1],
427  *(PolymorphicDowncast<const PadDescriptor*>(&descriptor)),
428  reasonIfUnsupported);
429  case LayerType::Permute:
430  return IsPermuteSupported(infos[0],
431  infos[1],
432  *(PolymorphicDowncast<const PermuteDescriptor*>(&descriptor)),
433  reasonIfUnsupported);
435  return IsPooling2dSupported(infos[0],
436  infos[1],
437  *(PolymorphicDowncast<const Pooling2dDescriptor*>(&descriptor)),
438  reasonIfUnsupported);
440  return IsPooling3dSupported(infos[0],
441  infos[1],
442  *(PolymorphicDowncast<const Pooling3dDescriptor*>(&descriptor)),
443  reasonIfUnsupported);
444  case LayerType::Prelu:
445  return IsPreluSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
446  case LayerType::QLstm:
447  return IsQLstmSupported(infos[0],
448  infos[1],
449  infos[2],
450  infos[3],
451  infos[4],
452  infos[5],
453  *(PolymorphicDowncast<const QLstmDescriptor*>(&descriptor)),
454  lstmParamsInfo.value(),
456  case LayerType::Quantize:
457  return IsQuantizeSupported(infos[0], infos[1], reasonIfUnsupported);
459  return IsQuantizedLstmSupported(infos[0],
460  infos[1],
461  infos[2],
462  infos[3],
463  infos[4],
464  quantizedLstmParamsInfo.value(),
466  case LayerType::Rank:
467  return true;
468  case LayerType::Reshape:
469  return IsReshapeSupported(infos[0],
470  infos[1],
471  *(PolymorphicDowncast<const ReshapeDescriptor*>(&descriptor)),
472  reasonIfUnsupported);
473  case LayerType::Resize:
474  return IsResizeSupported(infos[0],
475  infos[1],
476  *(PolymorphicDowncast<const ResizeDescriptor*>(&descriptor)),
477  reasonIfUnsupported);
478  case LayerType::Reduce:
479  return IsReduceSupported(infos[0],
480  infos[1],
481  *(PolymorphicDowncast<const ReduceDescriptor*>(&descriptor)),
482  reasonIfUnsupported);
483  case LayerType::Shape:
484  return LayerSupportBase::IsShapeSupported(infos[0],
485  infos[1],
486  reasonIfUnsupported);
487  case LayerType::Slice:
488  return IsSliceSupported(infos[0],
489  infos[1],
490  *(PolymorphicDowncast<const SliceDescriptor*>(&descriptor)),
491  reasonIfUnsupported);
492  case LayerType::Softmax:
493  return IsSoftmaxSupported(infos[0],
494  infos[1],
495  *(PolymorphicDowncast<const SoftmaxDescriptor*>(&descriptor)),
496  reasonIfUnsupported);
498  return IsSpaceToBatchNdSupported(infos[0],
499  infos[1],
500  *(PolymorphicDowncast<const SpaceToBatchNdDescriptor*>(&descriptor)),
501  reasonIfUnsupported);
503  return IsSpaceToDepthSupported(infos[0],
504  infos[1],
505  *(PolymorphicDowncast<const SpaceToDepthDescriptor*>(&descriptor)),
506  reasonIfUnsupported);
507  case LayerType::Splitter:
508  {
509  std::vector<TensorInfo> outputInfos;
510  for (uint32_t i = 1; i < infos.size(); i++)
511  {
512  outputInfos.push_back(infos[i]);
513  }
514  return IsSplitterSupported(infos[0],
515  {outputInfos.begin(), outputInfos.end()},
516  *(PolymorphicDowncast<const ViewsDescriptor*>(&descriptor)),
518  }
519  case LayerType::Stack:
520  {
521  std::vector<const TensorInfo*> inputInfos;
522  for (uint32_t i = 0; i < infos.size() - 1; i++)
523  {
524  inputInfos.push_back(&infos[i]);
525  }
526  return IsStackSupported(inputInfos,
527  infos[infos.size() - 1],
528  *(PolymorphicDowncast<const StackDescriptor*>(&descriptor)),
530  }
532  return IsStridedSliceSupported(infos[0],
533  infos[1],
534  *(PolymorphicDowncast<const StridedSliceDescriptor*>(&descriptor)),
535  reasonIfUnsupported);
537  return IsSubtractionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
539  return IsTransposeSupported(infos[0],
540  infos[1],
541  *(PolymorphicDowncast<const TransposeDescriptor*>(&descriptor)),
542  reasonIfUnsupported);
544  {
545  if (infos.size() != 4)
546  {
547  throw InvalidArgumentException("Invalid number of TransposeConvolution2d TensorInfos. "
548  "TensorInfos should be of format: {input, output, weights, biases}.");
549  }
550 
551  auto desc = *(PolymorphicDowncast<const TransposeConvolution2dDescriptor*>(&descriptor));
552  if (infos[3] == TensorInfo())
553  {
554  return IsTransposeConvolution2dSupported(infos[0],
555  infos[1],
556  desc,
557  infos[2],
558  EmptyOptional(),
559  reasonIfUnsupported);
560  }
561  else
562  {
563  return IsTransposeConvolution2dSupported(infos[0],
564  infos[1],
565  desc,
566  infos[2],
567  infos[3],
568  reasonIfUnsupported);
569  }
570  }
573  infos[1],
574  infos[2],
575  infos[3],
576  infos[4],
577  infos[5],
578  *(PolymorphicDowncast<const
579  UnidirectionalSequenceLstmDescriptor*>(&descriptor)),
580  lstmParamsInfo.value(),
582  case LayerType::Unmap:
583  return true;
584  default:
585  // layers not supported in neon by default:
586  // debug, fakequantization, precompiled,
587  // standin, switch
588  return false;
589  }
590 }
591 
593  const TensorInfo& output,
595  Optional<std::string&> reasonIfUnsupported) const
596 {
597  IgnoreUnused(descriptor);
599  reasonIfUnsupported,
600  input,
601  output,
602  descriptor);
603 }
604 
606  const TensorInfo& input1,
607  const TensorInfo& output,
608  Optional<std::string&> reasonIfUnsupported) const
609 {
611  reasonIfUnsupported,
612  input0,
613  input1,
614  output,
615  nullptr);
616 }
617 
619  const TensorInfo& output,
621  Optional<std::string&> reasonIfUnsupported) const
622 {
624  reasonIfUnsupported,
625  input,
626  output,
627  descriptor);
628 }
629 
631  const TensorInfo& output,
632  const TensorInfo& mean,
633  const TensorInfo& var,
634  const TensorInfo& beta,
635  const TensorInfo& gamma,
637  Optional<std::string&> reasonIfUnsupported) const
638 {
640  reasonIfUnsupported,
641  input,
642  output,
643  mean,
644  var,
645  beta,
646  gamma,
647  descriptor,
648  nullptr);
649 }
650 
652  const TensorInfo& output,
654  Optional<std::string&> reasonIfUnsupported) const
655 {
657  reasonIfUnsupported,
658  input,
659  output,
660  descriptor);
661 }
662 
664  const TensorInfo& output,
665  Optional<std::string&> reasonIfUnsupported) const
666 {
668  reasonIfUnsupported,
669  input,
670  output);
671 }
672 
674  const TensorInfo& output,
676  Optional<std::string&> reasonIfUnsupported) const
677 {
679  reasonIfUnsupported,
680  input,
681  output,
682  descriptor);
683 }
684 
686  const TensorInfo& input1,
687  const TensorInfo& output,
689  Optional<std::string&> reasonIfUnsupported) const
690 {
691 
693  reasonIfUnsupported,
694  input0,
695  input1,
696  output,
697  descriptor);
698 }
699 
700 bool NeonLayerSupport::IsConcatSupported(const std::vector<const TensorInfo*> inputs,
701  const TensorInfo& output,
703  Optional<std::string&> reasonIfUnsupported) const
704 {
705  if (descriptor.GetNumDimensions() <= descriptor.GetConcatAxis())
706  {
707  SetValueChecked(reasonIfUnsupported, "Neon Concat: Concat axis > Number of dimensions.");
708  return false;
709  }
710 
711  unsigned int concatInnerAxis = (descriptor.GetNumDimensions() - descriptor.GetConcatAxis()) - 1;
712  if(concatInnerAxis < 3) // Width, height, or channels
713  {
715  reasonIfUnsupported,
716  inputs,
717  output,
718  descriptor);
719  }
720  else if (concatInnerAxis == 3)
721  {
722  for (auto& input : inputs)
723  {
724  if (input && !output.IsTypeSpaceMatch(*input)) // Cannot use sub-tensors if the types are not same space
725  {
726  SetValueChecked(reasonIfUnsupported, "Neon Concat: Types and quantization parameters must match.");
727  return false;
728  }
729  }
730  return true; // Sub-tensors support concat along batch
731  }
732  else // > 4 dimensions not supported.
733  {
734  SetValueChecked(reasonIfUnsupported, "Neon Concat: Maximum of 4 dimensions supported.");
735  return false;
736  }
737 }
738 
740  Optional<std::string&> reasonIfUnsupported) const
741 {
743  reasonIfUnsupported,
744  output);
745 }
746 
748  const TensorInfo& output,
749  Optional<std::string&> reasonIfUnsupported) const
750 {
751  armnn::IgnoreUnused(input);
752  armnn::IgnoreUnused(output);
753  armnn::IgnoreUnused(reasonIfUnsupported);
754  return true;
755 }
756 
758  const TensorInfo& output,
759  Optional<std::string&> reasonIfUnsupported) const
760 {
761  armnn::IgnoreUnused(input);
762  armnn::IgnoreUnused(output);
763  armnn::IgnoreUnused(reasonIfUnsupported);
764  return true;
765 }
766 
768  const TensorInfo& output,
769  Optional<std::string&> reasonIfUnsupported) const
770 {
771  armnn::IgnoreUnused(input);
772  armnn::IgnoreUnused(output);
773  armnn::IgnoreUnused(reasonIfUnsupported);
774  return true;
775 }
776 
778  const TensorInfo& output,
779  Optional<std::string&> reasonIfUnsupported) const
780 {
781  armnn::IgnoreUnused(input);
782  armnn::IgnoreUnused(output);
783  armnn::IgnoreUnused(reasonIfUnsupported);
784  return true;
785 }
786 
788  const TensorInfo& output,
790  const TensorInfo& weights,
792  Optional<std::string&> reasonIfUnsupported) const
793 {
794  bool isFastMathEnabled = false;
795 #if defined(ARMCOMPUTENEON_ENABLED)
796  if (m_ModelContextPtr)
797  {
798  if (m_ModelContextPtr.get() != nullptr)
799  {
800  auto modelOptions = dynamic_cast<NeonBackendModelContext*>(m_ModelContextPtr.get());
801  if (modelOptions)
802  {
803  isFastMathEnabled = modelOptions->IsFastMathEnabled();
804  }
805  }
806  }
807 #endif
808 
810  reasonIfUnsupported,
811  input,
812  output,
813  descriptor,
814  weights,
815  biases,
816  isFastMathEnabled,
817  nullptr);
818 }
819 
821  const TensorInfo& output,
823  const TensorInfo& weights,
825  Optional<std::string&> reasonIfUnsupported) const
826 {
827  bool isFastMathEnabled = false;
828 #if defined(ARMCOMPUTENEON_ENABLED)
829  if (m_ModelContextPtr)
830  {
831  if (m_ModelContextPtr.get() != nullptr)
832  {
833  auto modelOptions = dynamic_cast<NeonBackendModelContext*>(m_ModelContextPtr.get());
834  if (modelOptions)
835  {
836  isFastMathEnabled = modelOptions->IsFastMathEnabled();
837  }
838  }
839  }
840 #endif
841 
843  reasonIfUnsupported,
844  input,
845  output,
846  descriptor,
847  weights,
848  biases,
849  isFastMathEnabled,
850  nullptr);
851 }
852 
854  const TensorInfo& output,
856  Optional<std::string&> reasonIfUnsupported) const
857 {
859  reasonIfUnsupported,
860  input,
861  output,
862  descriptor);
863 }
864 
866  const TensorInfo& output,
868  const TensorInfo& weights,
870  Optional<std::string&> reasonIfUnsupported) const
871 {
873  reasonIfUnsupported,
874  input,
875  output,
876  descriptor,
877  weights,
878  biases,
879  nullptr);
880 }
881 
883  const TensorInfo& output,
884  Optional<std::string&> reasonIfUnsupported) const
885 {
887  reasonIfUnsupported,
888  input,
889  output);
890 }
891 
893  const TensorInfo& output,
895  const TensorInfo& weights,
897  Optional<std::string&> reasonIfUnsupported) const
898 {
900  reasonIfUnsupported,
901  input,
902  output,
903  descriptor,
904  weights,
905  biases,
906  nullptr);
907 }
908 
910  const TensorInfo& output,
912  Optional<std::string&> reasonIfUnsupported) const
913 {
914  switch(descriptor.m_Operation)
915  {
916  case UnaryOperation::Abs:
918  reasonIfUnsupported,
919  input,
920  output);
921  case UnaryOperation::Exp:
923  reasonIfUnsupported,
924  input,
925  output);
928  reasonIfUnsupported,
929  input,
930  output);
931  case UnaryOperation::Log:
933  reasonIfUnsupported,
934  input,
935  output);
936  case UnaryOperation::Neg:
938  reasonIfUnsupported,
939  input,
940  output);
943  reasonIfUnsupported,
944  input,
945  output);
946  case UnaryOperation::Sin:
948  reasonIfUnsupported,
949  input,
950  output);
953  reasonIfUnsupported,
954  input,
955  output);
956  default:
957  return false;
958  }
959 }
960 
962  const TensorInfo& output,
963  const FillDescriptor& descriptor,
964  Optional<std::string&> reasonIfUnsupported) const
965 {
966  armnn::IgnoreUnused(input);
967  armnn::IgnoreUnused(output);
968  armnn::IgnoreUnused(descriptor);
969 
970  return IsNeonBackendSupported(reasonIfUnsupported);
971 }
972 
974  const TensorInfo& output,
975  Optional<std::string&> reasonIfUnsupported) const
976 {
977  armnn::IgnoreUnused(output);
978  return IsNeonBackendSupported(reasonIfUnsupported) &&
979  IsSupportedForDataTypeGeneric(reasonIfUnsupported,
980  input.GetDataType(),
981  &FalseFuncF16<>,
982  &TrueFunc<>,
983  &FalseFuncU8<>,
984  &FalseFuncI32<>,
985  &FalseFuncU8<>);
986 }
987 
989  const TensorInfo& output,
990  const TensorInfo& weights,
991  const TensorInfo& biases,
993  Optional<std::string&> reasonIfUnsupported) const
994 {
996  reasonIfUnsupported,
997  input,
998  output,
999  weights,
1000  biases,
1001  descriptor,
1002  nullptr);
1003 }
1004 
1006  const TensorInfo& input1,
1007  const TensorInfo& output,
1009  Optional<std::string&> reasonIfUnsupported) const
1010 {
1012  reasonIfUnsupported,
1013  input0,
1014  input1,
1015  output,
1016  descriptor);
1017 }
1018 
1020  const TensorInfo& input1,
1021  const TensorInfo& output,
1022  Optional<std::string&> reasonIfUnsupported) const
1023 {
1025  reasonIfUnsupported,
1026  input0,
1027  input1,
1028  output);
1029 }
1030 
1032  Optional<std::string&> reasonIfUnsupported) const
1033 {
1034  return IsNeonBackendSupported(reasonIfUnsupported, input);
1035 }
1036 
1038  const TensorInfo& output,
1040  Optional<std::string&> reasonIfUnsupported) const
1041 {
1043  reasonIfUnsupported,
1044  input,
1045  output,
1046  descriptor);
1047 }
1048 
1050  const TensorInfo& output,
1052  Optional<std::string&> reasonIfUnsupported) const
1053 {
1054  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonL2NormalizationWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1055 }
1056 
1058  const TensorInfo& input1,
1059  const TensorInfo& output,
1061  Optional<std::string&> reasonIfUnsupported) const
1062 {
1063  switch(descriptor.m_Operation)
1064  {
1067  reasonIfUnsupported,
1068  input0,
1069  input1,
1070  output);
1073  reasonIfUnsupported,
1074  input0,
1075  input1,
1076  output);
1077  default:
1078  return false;
1079  }
1080 }
1081 
1083  const TensorInfo& output,
1085  Optional<std::string&> reasonIfUnsupported) const
1086 {
1087  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonLogSoftmaxWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1088 }
1089 
1091  const TensorInfo& outputStateIn,
1092  const TensorInfo& cellStateIn,
1093  const TensorInfo& scratchBuffer,
1094  const TensorInfo& outputStateOut,
1095  const TensorInfo& cellStateOut,
1096  const TensorInfo& output,
1097  const LstmDescriptor& descriptor,
1099  Optional<std::string&> reasonIfUnsupported) const
1100 {
1102  reasonIfUnsupported,
1103  input,
1104  outputStateIn,
1105  cellStateIn,
1106  scratchBuffer,
1107  outputStateOut,
1108  cellStateOut,
1109  output,
1110  descriptor,
1111  paramsInfo);
1112 }
1113 
1115  const TensorInfo& input1,
1116  const TensorInfo& output,
1117  Optional<std::string&> reasonIfUnsupported) const
1118 {
1120  reasonIfUnsupported,
1121  input0,
1122  input1,
1123  output);
1124 }
1125 
1127  const TensorInfo& output,
1128  const MeanDescriptor& descriptor,
1129  Optional<std::string&> reasonIfUnsupported) const
1130 {
1132  reasonIfUnsupported,
1133  input,
1134  output,
1135  descriptor);
1136 }
1137 
1139  const TensorInfo& input1,
1140  const TensorInfo& output,
1141  Optional<std::string&> reasonIfUnsupported) const
1142 {
1144  reasonIfUnsupported,
1145  input0,
1146  input1,
1147  output);
1148 }
1149 
1151  const TensorInfo& input1,
1152  const TensorInfo& output,
1153  Optional<std::string&> reasonIfUnsupported) const
1154 {
1156  reasonIfUnsupported,
1157  input0,
1158  input1,
1159  output,
1160  nullptr);
1161 }
1162 
1164  const TensorInfo& input1,
1165  const TensorInfo& output,
1166  Optional<std::string&> reasonIfUnsupported) const
1167 {
1169  reasonIfUnsupported,
1170  input0,
1171  input1,
1172  output,
1173  nullptr);
1174 }
1175 
1177  const TensorInfo& output,
1179  Optional<std::string&> reasonIfUnsupported) const
1180 {
1182  reasonIfUnsupported,
1183  input,
1184  output,
1185  descriptor);
1186 }
1187 
1189  Optional<std::string&> reasonIfUnsupported) const
1190 {
1191  return IsNeonBackendSupported(reasonIfUnsupported, output);
1192 }
1193 
1195  const TensorInfo& output,
1196  const PadDescriptor& descriptor,
1197  Optional<std::string&> reasonIfUnsupported) const
1198 {
1200  reasonIfUnsupported,
1201  input,
1202  output,
1203  descriptor);
1204 }
1205 
1207  const TensorInfo& output,
1209  Optional<std::string&> reasonIfUnsupported) const
1210 {
1211  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPermuteWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1212 }
1213 
1215  const TensorInfo& output,
1217  Optional<std::string&> reasonIfUnsupported) const
1218 {
1219  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPooling2dWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1220 }
1221 
1223  const TensorInfo& output,
1225  Optional<std::string&> reasonIfUnsupported) const
1226 {
1227  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPooling3dWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1228 }
1229 
1231  const armnn::TensorInfo &alpha,
1232  const armnn::TensorInfo &output,
1233  armnn::Optional<std::string &> reasonIfUnsupported) const
1234 {
1235  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPreluWorkloadValidate, reasonIfUnsupported, input, alpha, output);
1236 }
1237 
1241  const TensorInfo& outputStateOut,
1242  const TensorInfo& cellStateOut,
1243  const TensorInfo& output,
1244  const QLstmDescriptor& descriptor,
1246  Optional<std::string&> reasonIfUnsupported) const
1247 {
1248  // Check required here in order to pass IsLayerSupported for datatypes tests
1249  if (input.GetDataType() == armnn::DataType::QAsymmS8 &&
1250  previousOutputIn.GetDataType() == armnn::DataType::QAsymmS8 &&
1251  previousCellStateIn.GetDataType() == armnn::DataType::QSymmS16 &&
1252  outputStateOut.GetDataType() == armnn::DataType::QAsymmS8 &&
1253  cellStateOut.GetDataType() == armnn::DataType::QSymmS16 &&
1255  {
1257  reasonIfUnsupported,
1258  input,
1259  previousCellStateIn,
1260  previousOutputIn,
1261  cellStateOut,
1262  outputStateOut,
1263  output,
1264  descriptor,
1265  paramsInfo);
1266  }
1267  else
1268  {
1269  return false;
1270  }
1271 }
1272 
1274  const TensorInfo& output,
1275  Optional<std::string&> reasonIfUnsupported) const
1276 {
1278  reasonIfUnsupported,
1279  input,
1280  output);
1281 }
1282 
1284  const TensorInfo& cellStateIn,
1285  const TensorInfo& outputStateIn,
1286  const TensorInfo& cellStateOut,
1287  const TensorInfo& outputStateOut,
1289  Optional<std::string&> reasonIfUnsupported) const
1290 {
1292  reasonIfUnsupported,
1293  input,
1294  cellStateIn,
1295  outputStateIn,
1296  cellStateOut,
1297  outputStateOut,
1298  paramsInfo);
1299 }
1300 
1302  const TensorInfo& output,
1304  Optional<std::string&> reasonIfUnsupported) const
1305 {
1307  reasonIfUnsupported,
1308  input,
1309  output,
1310  descriptor);
1311 }
1312 
1314  const TensorInfo& output,
1316  Optional<std::string&> reasonIfUnsupported) const
1317 {
1318  armnn::IgnoreUnused(descriptor);
1320  reasonIfUnsupported,
1321  input,
1322  output);
1323 }
1324 
1326  const TensorInfo& output,
1328  Optional<std::string&> reasonIfUnsupported) const
1329 {
1331  reasonIfUnsupported,
1332  input,
1333  output,
1334  descriptor);
1335 }
1336 
1338  const TensorInfo& output,
1339  const SliceDescriptor& descriptor,
1340  Optional<std::string&> reasonIfUnsupported) const
1341 {
1343  reasonIfUnsupported,
1344  input,
1345  output,
1346  descriptor);
1347 }
1348 
1350  const TensorInfo& output,
1352  Optional<std::string&> reasonIfUnsupported) const
1353 {
1354  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSoftmaxWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1355 }
1356 
1358  const TensorInfo& output,
1360  Optional<std::string&> reasonIfUnsupported) const
1361 {
1363  reasonIfUnsupported,
1364  input,
1365  output,
1366  descriptor);
1367 }
1368 
1370  const TensorInfo& output,
1372  Optional<std::string&> reasonIfUnsupported) const
1373 {
1375  reasonIfUnsupported,
1376  input,
1377  output,
1378  descriptor);
1379 }
1380 
1382  const std::vector<std::reference_wrapper<TensorInfo>>& outputs,
1383  const ViewsDescriptor& descriptor,
1384  Optional<std::string&> reasonIfUnsupported) const
1385 {
1386 #if defined(ARMCOMPUTENEON_ENABLED)
1387  // Split along the last dimension, cannot use sub-tensors
1388  // as width and height of the sub-tensors do not match
1389  // the width and height of the parent tensor
1390  // in case of input with more than 2D.
1391  std::set<unsigned int> splitAxis = ComputeSplitAxis(descriptor, input.GetShape());
1392  if (descriptor.GetNumDimensions() > 2 && splitAxis.size() == 1 &&
1393  *splitAxis.begin() == descriptor.GetNumDimensions() - 1 )
1394  {
1396  reasonIfUnsupported,
1397  input,
1398  outputs,
1399  *splitAxis.begin());
1400  }
1401 #endif
1402  IgnoreUnused(descriptor);
1403  for (auto output : outputs)
1404  {
1405  if (!input.IsTypeSpaceMatch(output)) // Cannot use sub-tensors if the types are not same space
1406  {
1407  SetValueChecked(reasonIfUnsupported, "Neon Splitter: Types and quantization parameters must match.");
1408  return false;
1409  }
1410  }
1411  return true;
1412 }
1413 
1414 bool NeonLayerSupport::IsStackSupported(const std::vector<const TensorInfo*>& inputs,
1415  const TensorInfo& output,
1416  const StackDescriptor& descriptor,
1417  Optional<std::string&> reasonIfUnsupported) const
1418 {
1420  reasonIfUnsupported,
1421  inputs,
1422  output,
1423  descriptor);
1424 }
1425 
1427  const TensorInfo& output,
1429  Optional<std::string&> reasonIfUnsupported) const
1430 {
1432  reasonIfUnsupported,
1433  input,
1434  output,
1435  descriptor);
1436 }
1437 
1439  const TensorInfo& input1,
1440  const TensorInfo& output,
1441  Optional<std::string&> reasonIfUnsupported) const
1442 {
1444  reasonIfUnsupported,
1445  input0,
1446  input1,
1447  output,
1448  nullptr);
1449 }
1450 
1452  const TensorInfo& output,
1454  const TensorInfo& weights,
1456  Optional<std::string&> reasonIfUnsupported) const
1457 {
1459  reasonIfUnsupported,
1460  input,
1461  output,
1462  descriptor,
1463  weights,
1464  biases);
1465 }
1466 
1468  const TensorInfo& output,
1470  Optional<std::string&> reasonIfUnsupported) const
1471 {
1472  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonTransposeWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
1473 }
1474 
1476  const TensorInfo& outputStateIn,
1477  const TensorInfo& cellStateIn,
1478  const TensorInfo& outputStateOut,
1479  const TensorInfo& cellStateOut,
1480  const TensorInfo& output,
1483  Optional<std::string&> reasonIfUnsupported) const
1484 {
1485  if (input.GetDataType() == armnn::DataType::QAsymmS8 &&
1486  outputStateIn.GetDataType() == armnn::DataType::QAsymmS8 &&
1487  cellStateIn.GetDataType() == armnn::DataType::QSymmS16 &&
1488  outputStateOut.GetDataType() == armnn::DataType::QAsymmS8 &&
1489  cellStateOut.GetDataType() == armnn::DataType::QSymmS16 &&
1491  {
1493  reasonIfUnsupported,
1494  input,
1495  outputStateIn,
1496  cellStateIn,
1497  outputStateOut,
1498  cellStateOut,
1499  output,
1500  descriptor,
1501  paramsInfo);
1502  }
1503  else
1504  {
1506  reasonIfUnsupported,
1507  input,
1508  outputStateIn,
1509  cellStateIn,
1510  outputStateOut,
1511  cellStateOut,
1512  output,
1513  descriptor,
1514  paramsInfo);
1515  }
1516 }
1517 
1518 } // namespace armnn
arm_compute::Status NeonGatherWorkloadValidate(const TensorInfo &input, const TensorInfo &indices, const TensorInfo &output, const GatherDescriptor &descriptor)
arm_compute::Status NeonNegWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsConcatSupported(const std::vector< const TensorInfo *> inputs, const TensorInfo &output, const OriginsDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
UnaryOperation m_Operation
Specifies the elementwiseUnary operation to execute.
bool IsSliceSupported(const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsSoftmaxSupported(const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsConvolution2dSupported(const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A ViewsDescriptor for the SplitterLayer.
bool IsTypeSpaceMatch(const TensorInfo &other) const
Check that the types are the same and, if quantize, that the quantization parameters are the same...
Definition: Tensor.cpp:432
bool IsConvertFp32ToFp16Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonSpaceToDepthWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor)
arm_compute::Status NeonSplitterWorkloadValidate(const TensorInfo &input, const std::vector< std::reference_wrapper< TensorInfo >> &outputs, unsigned int splitAxis)
A TransposeConvolution2dDescriptor for the TransposeConvolution2dLayer.
const TensorInfo const TensorInfo const TensorInfo const TensorInfo const TensorInfo const TensorInfo const LstmDescriptor const LstmInputParamsInfo & paramsInfo
const TensorShape & GetShape() const
Definition: Tensor.hpp:191
const TensorInfo & output
arm_compute::Status NeonLogSoftmaxWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor)
A ReshapeDescriptor for the ReshapeLayer.
bool IsGatherSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const GatherDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported) const override
arm_compute::Status NeonBatchNormalizationValidate(const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, const ActivationDescriptor *activationDescriptor)
bool IsDetectionPostProcessSupported(const TensorInfo &boxEncodings, const TensorInfo &scores, const TensorInfo &anchors, const TensorInfo &detectionBoxes, const TensorInfo &detectionClasses, const TensorInfo &detectionScores, const TensorInfo &numDetections, const DetectionPostProcessDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsMemImportSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsDilatedDepthwiseConvolutionSupported(const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string &> reason=EmptyOptional()) const override
bool IsQuantizedLstmSupported(const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const QuantizedLstmInputParamsInfo &paramsInfo, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
uint32_t GetNumDimensions() const
Get the number of dimensions.
A ComparisonDescriptor for the ComparisonLayer.
Definition: Descriptors.hpp:89
bool IsDepthToSpaceSupported(const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
const TensorInfo const TensorInfo const TensorInfo const TensorInfo const TensorInfo & gamma
bool IsL2NormalizationSupported(const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsCastSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsStridedSliceSupported(const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
const std::vector< std::reference_wrapper< TensorInfo > > & outputs
bool IsLayerSupported(const LayerType &type, const std::vector< TensorInfo > &infos, const BaseDescriptor &descriptor, const Optional< LstmInputParamsInfo > &lstmParamsInfo, const Optional< QuantizedLstmInputParamsInfo > &quantizedLstmParamsInfo, Optional< std::string &> reasonIfUnsupported) const override
arm_compute::Status NeonGatherNdWorkloadValidate(const TensorInfo &paramsInfo, const TensorInfo &indicesInfo, const TensorInfo &outputInfo)
bool IsConvertFp32ToBf16Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsAdditionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A Convolution2dDescriptor for the Convolution2dLayer.
bool IsGatherNdSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported) const
arm_compute::Status NeonDepthwiseConvolutionWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, const ActivationDescriptor *activationDescriptor)
arm_compute::Status NeonActivationWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor)
arm_compute::Status NeonMinimumWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output)
Validate function for validating the inputs and output.
arm_compute::Status NeonStridedSliceWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor)
const TensorInfo const ActivationDescriptor Optional< std::string & > reasonIfUnsupported
arm_compute::Status NeonTransposeConvolution2dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases)
arm_compute::Status NeonPooling3dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const Pooling3dDescriptor &descriptor)
arm_compute::Status NeonLstmFloatWorkloadValidate(const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo)
bool IsShapeSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonQLstmWorkloadValidate(const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo)
arm_compute::Status NeonSliceWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor)
arm_compute::Status NeonFullyConnectedWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const Optional< TensorInfo > &biases, const FullyConnectedDescriptor &descriptor, const ActivationDescriptor *activationDescriptor)
A LogicalBinaryDescriptor for the LogicalBinaryLayer.
bool IsConstantSupported(const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
Copyright (c) 2021 ARM Limited and Contributors.
arm_compute::Status NeonQuantizeWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsPadSupported(const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
void IgnoreUnused(Ts &&...)
const TensorInfo const ActivationDescriptor & descriptor
std::set< unsigned int > ComputeSplitAxis(const armnn::SplitterDescriptor &desc, const TensorShape &input)
arm_compute::Status NeonAdditionWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor)
A SpaceToDepthDescriptor for the SpaceToDepthLayer.
arm_compute::Status NeonLogWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
arm_compute::Status NeonLogicalAndWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output)
arm_compute::Status NeonInstanceNormalizationWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor)
bool IsPooling2dSupported(const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonLogicalOrWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output)
LogicalBinaryOperation m_Operation
Specifies the logical operation to execute.
A BatchToSpaceNdDescriptor for the BatchToSpaceNdLayer.
bool IsActivationSupported(const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
The NeonBackendModelContext is used to pass in Neon specific backend ModelOptions.
bool IsMultiplicationSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
const TensorInfo & outputStateIn
const TensorInfo const TensorInfo & previousCellStateIn
bool IsComparisonSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A ResizeBilinearDescriptor for the ResizeBilinearLayer.
const TensorInfo & alpha
arm_compute::Status NeonL2NormalizationWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor)
arm_compute::Status NeonAbsWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
Base class for all descriptors.
Definition: Descriptors.hpp:22
A StackDescriptor for the StackLayer.
arm_compute::Status NeonQuantizedLstmWorkloadValidate(const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const QuantizedLstmInputParamsInfo &paramsInfo)
arm_compute::Status NeonStackWorkloadValidate(const std::vector< const TensorInfo *> &inputs, const TensorInfo &output, const StackDescriptor &descriptor)
arm_compute::Status NeonMeanWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &descriptor)
arm_compute::Status NeonReduceWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &descriptor)
arm_compute::Status NeonSpaceToBatchNdWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor)
arm_compute::Status NeonSubtractionWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor)
A PadDescriptor for the PadLayer.
bool IsLogicalBinarySupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const LogicalBinaryDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported) const override
bool IsConvolution3dSupported(const TensorInfo &input, const TensorInfo &output, const Convolution3dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
DataType
Definition: Types.hpp:48
const TensorInfo const TensorInfo & cellStateIn
bool IsFillSupported(const TensorInfo &input, const TensorInfo &output, const FillDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
An LstmDescriptor for the LstmLayer.
arm_compute::Status NeonExpWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
arm_compute::Status NeonReshapeWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsInputSupported(const TensorInfo &input, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonBatchToSpaceNdWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &descriptor)
bool IsStackSupported(const std::vector< const TensorInfo *> &inputs, const TensorInfo &output, const StackDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
std::shared_ptr< IBackendModelContext > IBackendSpecificModelContextPtr
A L2NormalizationDescriptor for the L2NormalizationLayer.
const TensorInfo const TensorInfo const TensorInfo & var
bool IsMinimumSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsPreluSupported(const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsOutputSupported(const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
An ArgMinMaxDescriptor for ArgMinMaxLayer.
Definition: Descriptors.hpp:67
bool IsInstanceNormalizationSupported(const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
DataType GetDataType() const
Definition: Tensor.hpp:198
An OriginsDescriptor for the ConcatLayer.
A ReduceDescriptor for the REDUCE operators.
A FullyConnectedDescriptor for the FullyConnectedLayer.
arm_compute::Status NeonLogicalNotWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
const TensorInfo const TensorInfo const TensorInfo const TensorInfo & outputStateOut
bool IsNormalizationSupported(const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsElementwiseUnarySupported(const TensorInfo &input, const TensorInfo &output, const ElementwiseUnaryDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
const TensorInfo const TensorInfo const TensorInfo const TensorInfo const TensorInfo & cellStateOut
arm_compute::Status NeonCastValidate(const TensorInfo &input, const TensorInfo &output)
A GatherDescriptor for the GatherLayer.
Status
enumeration
Definition: Types.hpp:42
arm_compute::Status NeonChannelShuffleValidate(const TensorInfo &input, const TensorInfo &output, const ChannelShuffleDescriptor &descriptor)
arm_compute::Status NeonComparisonWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor)
bool IsDivisionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonUnidirectionalSequenceLstmWorkloadValidate(const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo)
const TensorInfo const TensorInfo const TensorInfo const TensorInfo & beta
arm_compute::Status NeonConcatWorkloadValidate(const std::vector< const TensorInfo *> &inputs, const TensorInfo &output, const OriginsDescriptor &descriptor)
arm_compute::Status NeonPermuteWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor)
bool IsFloorSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A QLstmDescriptor for the QLstmLayer.
bool IsMergeSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
DestType PolymorphicDowncast(SourceType *value)
Polymorphic downcast for build in pointers only.
arm_compute::Status NeonConvolution2dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor)
bool IsLogSoftmaxSupported(const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
An ActivationDescriptor for the ActivationLayer.
Definition: Descriptors.hpp:36
bool IsResizeSupported(const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsMemCopySupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsTransposeConvolution2dSupported(const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A SliceDescriptor for the SliceLayer.
A Convolution3dDescriptor for the Convolution3dLayer.
arm_compute::Status NeonDequantizeWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsSplitterSupported(const TensorInfo &input, const std::vector< std::reference_wrapper< TensorInfo >> &outputs, const ViewsDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
const TensorInfo & previousOutputIn
bool IsReshapeSupported(const TensorInfo &input, const TensorInfo &output, const ReshapeDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonDivisionWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor)
bool IsBatchNormalizationSupported(const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A Pooling3dDescriptor for the Pooling3dLayer.
arm_compute::Status NeonConvolution3dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const Convolution3dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor)
void SetValueChecked(Optional< T &> optionalRef, V &&val)
A SpaceToBatchNdDescriptor for the SpaceToBatchNdLayer.
arm_compute::Status NeonNormalizationWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor)
arm_compute::Status NeonUnidirectionalSequenceLstmFloatWorkloadValidate(const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo)
bool IsQuantizeSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
EmptyOptional is used to initialize the Optional class in case we want to have default value for an O...
Definition: Optional.hpp:32
arm_compute::Status NeonRsqrtWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsConvertFp16ToFp32Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonPadWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor)
A ElementwiseUnaryDescriptor for the ElementwiseUnaryLayer.
bool IsSubtractionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsLstmSupported(const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonArgMinMaxWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor)
arm_compute::Status NeonSoftmaxWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor)
bool IsSpaceToBatchNdSupported(const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonTransposeWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor)
const TensorInfo const Convolution2dDescriptor const TensorInfo const Optional< TensorInfo > & biases
bool IsPooling3dSupported(const TensorInfo &input, const TensorInfo &output, const Pooling3dDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
uint32_t GetNumDimensions() const
Get the number of dimensions.
arm_compute::Status NeonSqrtWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
A MeanDescriptor for the MeanLayer.
bool IsDepthwiseConvolutionSupported(const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonMaximumWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output)
bool IsArgMinMaxSupported(const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsSpaceToDepthSupported(const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A TransposeDescriptor for the TransposeLayer.
A StridedSliceDescriptor for the StridedSliceLayer.
bool IsTransposeSupported(const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsMaximumSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonConstantWorkloadValidate(const TensorInfo &output)
const TensorInfo & input1
bool IsFullyConnectedSupported(const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const TensorInfo &biases, const FullyConnectedDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsMeanSupported(const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsBatchToSpaceNdSupported(const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonPreluWorkloadValidate(const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output)
#define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported,...)
arm_compute::Status NeonDepthToSpaceWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor)
A Pooling2dDescriptor for the Pooling2dLayer.
arm_compute::Status NeonResizeWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor)
bool IsUnidirectionalSequenceLstmSupported(const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const UnidirectionalSequenceLstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo, Optional< std::string &> reasonIfUnsupported) const override
A NormalizationDescriptor for the NormalizationLayer.
bool IsQLstmSupported(const TensorInfo &input, const TensorInfo &previousOutputIn, const TensorInfo &previousCellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsConvertBf16ToFp32Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
const TensorInfo const TensorInfo const TensorInfo & scratchBuffer
An InstanceNormalizationDescriptor for InstanceNormalizationLayer.
unsigned int GetConcatAxis() const
Get the concatenation axis value.
A ChannelShuffleDescriptor for the ChannelShuffle operator.
bool IsSupportedForDataTypeGeneric(Optional< std::string &> reasonIfUnsupported, DataType dataType, Float16Func float16FuncPtr, Float32Func float32FuncPtr, Uint8Func uint8FuncPtr, Int32Func int32FuncPtr, BooleanFunc booleanFuncPtr, Params &&... params)
A SoftmaxDescriptor for the SoftmaxLayer.
bool IsReduceSupported(const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
const TensorInfo const Convolution2dDescriptor const TensorInfo & weights
arm_compute::Status NeonPooling2dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor)
arm_compute::Status NeonSinWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsDequantizeSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A DepthwiseConvolution2dDescriptor for the DepthwiseConvolution2dLayer.
A FillDescriptor for the FillLayer.
A BatchNormalizationDescriptor for the BatchNormalizationLayer.
arm_compute::Status NeonMultiplicationWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor)
bool IsChannelShuffleSupported(const TensorInfo &input, const TensorInfo &output, const ChannelShuffleDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
const TensorInfo const TensorInfo & mean
A PermuteDescriptor for the PermuteLayer.
LayerType
When adding a new layer, adapt also the LastLayer enum value in the enum class LayerType below...
Definition: Types.hpp:467
bool IsPermuteSupported(const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override