ArmNN
 21.02
NeonLayerSupport.cpp
Go to the documentation of this file.
1 //
2 // Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
3 // SPDX-License-Identifier: MIT
4 //
5 
6 #include "NeonLayerSupport.hpp"
7 #include "NeonBackendId.hpp"
9 
10 #include <armnn/Descriptors.hpp>
11 #include <armnn/Exceptions.hpp>
12 #include <armnn/Tensor.hpp>
13 #include <armnn/Types.hpp>
15 
16 #include <InternalTypes.hpp>
17 #include <LayerSupportCommon.hpp>
20 
21 #if defined(ARMCOMPUTENEON_ENABLED)
75 #endif
76 
77 namespace armnn
78 {
79 
80 namespace
81 {
82 
83 template< typename ... Args>
84 bool IsNeonBackendSupported(Optional<std::string&> reasonIfUnsupported, Args... args)
85 {
86  IgnoreUnused(reasonIfUnsupported, (args)...);
87 #if defined(ARMCOMPUTENEON_ENABLED)
88  return true;
89 #else
90  SetValueChecked(reasonIfUnsupported, "The armnn library has been built without NEON support");
91  return false;
92 #endif
93 }
94 
95 template<typename FloatFunc, typename Uint8Func, typename ... Params>
96 bool IsSupportedForDataTypeNeon(Optional<std::string&> reasonIfUnsupported,
97  DataType dataType,
98  FloatFunc floatFuncPtr,
99  Uint8Func uint8FuncPtr,
100  Params&&... params)
101 {
102  return IsNeonBackendSupported(reasonIfUnsupported) &&
103  IsSupportedForDataTypeGeneric(reasonIfUnsupported,
104  dataType,
105  floatFuncPtr,
106  floatFuncPtr,
107  uint8FuncPtr,
108  &FalseFunc<>,
109  &FalseFunc<>,
110  std::forward<Params>(params)...);
111 }
112 
113 #if defined(ARMCOMPUTENEON_ENABLED)
114 template<class FuncType, class... Args>
115 inline bool IsWorkloadSupported(FuncType& func, Optional<std::string&> reasonIfUnsupported, Args&&... args)
116 {
117  arm_compute::Status aclStatus = func(std::forward<Args>(args)...);
118  const bool supported = (aclStatus.error_code() == arm_compute::ErrorCode::OK);
119  if (!supported && reasonIfUnsupported)
120  {
121  reasonIfUnsupported.value() = aclStatus.error_description();
122  }
123  return supported;
124 }
125 
126 #define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
127  return IsWorkloadSupported(func, reasonIfUnsupported, __VA_ARGS__);
128 #else
129 #define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported, ...) \
130  return IsNeonBackendSupported(reasonIfUnsupported, __VA_ARGS__);
131 #endif
132 } // anonymous namespace
133 
135  : m_ModelContextPtr(modelContextPtr)
136 {
137 }
138 
140  : m_ModelContextPtr(nullptr)
141 {
142 }
143 
145  const TensorInfo& output,
146  Optional<std::string&> reasonIfUnsupported) const
147 {
149  return IsElementwiseUnarySupported(input, output, descriptor, reasonIfUnsupported);
150 }
151 
153  const TensorInfo& output,
154  const ActivationDescriptor& descriptor,
155  Optional<std::string&> reasonIfUnsupported) const
156 {
157  IgnoreUnused(descriptor);
159  reasonIfUnsupported,
160  input,
161  output,
162  descriptor);
163 }
164 
166  const TensorInfo& input1,
167  const TensorInfo& output,
168  Optional<std::string&> reasonIfUnsupported) const
169 {
171  reasonIfUnsupported,
172  input0,
173  input1,
174  output,
175  nullptr);
176 }
177 
179  const TensorInfo& output,
180  const ArgMinMaxDescriptor& descriptor,
181  Optional<std::string&> reasonIfUnsupported) const
182 {
184  reasonIfUnsupported,
185  input,
186  output,
187  descriptor);
188 }
189 
191  const TensorInfo& output,
192  const TensorInfo& mean,
193  const TensorInfo& var,
194  const TensorInfo& beta,
195  const TensorInfo& gamma,
196  const BatchNormalizationDescriptor& descriptor,
197  Optional<std::string&> reasonIfUnsupported) const
198 {
200  reasonIfUnsupported,
201  input,
202  output,
203  mean,
204  var,
205  beta,
206  gamma,
207  descriptor,
208  nullptr);
209 }
210 
212  const TensorInfo& output,
213  const BatchToSpaceNdDescriptor& descriptor,
214  Optional<std::string&> reasonIfUnsupported) const
215 {
217  reasonIfUnsupported,
218  input,
219  output,
220  descriptor);
221 }
222 
224  const TensorInfo& input1,
225  const TensorInfo& output,
226  const ComparisonDescriptor& descriptor,
227  Optional<std::string&> reasonIfUnsupported) const
228 {
229 
231  reasonIfUnsupported,
232  input0,
233  input1,
234  output,
235  descriptor);
236 }
237 
238 bool NeonLayerSupport::IsConcatSupported(const std::vector<const TensorInfo*> inputs,
239  const TensorInfo& output,
240  const ConcatDescriptor& descriptor,
241  Optional<std::string&> reasonIfUnsupported) const
242 {
243  if (descriptor.GetNumDimensions() <= descriptor.GetConcatAxis())
244  {
245  SetValueChecked(reasonIfUnsupported, "Neon Concat: Concat axis > Number of dimensions.");
246  return false;
247  }
248 
249  unsigned int concatInnerAxis = (descriptor.GetNumDimensions() - descriptor.GetConcatAxis()) - 1;
250  if(concatInnerAxis < 3) // Width, height, or channels
251  {
253  reasonIfUnsupported,
254  inputs,
255  output,
256  descriptor);
257  }
258  else if (concatInnerAxis == 3)
259  {
260  for (auto& input : inputs)
261  {
262  if (input && !output.IsTypeSpaceMatch(*input)) // Cannot use sub-tensors if the types are not same space
263  {
264  SetValueChecked(reasonIfUnsupported, "Neon Concat: Types and quantization parameters must match.");
265  return false;
266  }
267  }
268  return true; // Sub-tensors support concat along batch
269  }
270  else // > 4 dimensions not supported.
271  {
272  SetValueChecked(reasonIfUnsupported, "Neon Concat: Maximum of 4 dimensions supported.");
273  return false;
274  }
275 }
276 
278  Optional<std::string&> reasonIfUnsupported) const
279 {
281  reasonIfUnsupported,
282  output);
283 }
284 
286  const TensorInfo& output,
287  Optional<std::string&> reasonIfUnsupported) const
288 {
289  armnn::IgnoreUnused(input);
290  armnn::IgnoreUnused(output);
291  armnn::IgnoreUnused(reasonIfUnsupported);
292  return true;
293 }
294 
296  const TensorInfo& output,
297  Optional<std::string&> reasonIfUnsupported) const
298 {
299  armnn::IgnoreUnused(input);
300  armnn::IgnoreUnused(output);
301  armnn::IgnoreUnused(reasonIfUnsupported);
302  return true;
303 }
304 
306  const TensorInfo& output,
307  Optional<std::string&> reasonIfUnsupported) const
308 {
309  armnn::IgnoreUnused(input);
310  armnn::IgnoreUnused(output);
311  armnn::IgnoreUnused(reasonIfUnsupported);
312  return true;
313 }
314 
316  const TensorInfo& output,
317  Optional<std::string&> reasonIfUnsupported) const
318 {
319  armnn::IgnoreUnused(input);
320  armnn::IgnoreUnused(output);
321  armnn::IgnoreUnused(reasonIfUnsupported);
322  return true;
323 }
324 
326  const TensorInfo& output,
327  const Convolution2dDescriptor& descriptor,
328  const TensorInfo& weights,
329  const Optional<TensorInfo>& biases,
330  Optional<std::string&> reasonIfUnsupported) const
331 {
332  bool isFastMathEnabled = false;
333 #if defined(ARMCOMPUTENEON_ENABLED)
334  if (m_ModelContextPtr)
335  {
336  if (m_ModelContextPtr.get() != nullptr)
337  {
338  auto modelOptions = dynamic_cast<NeonBackendModelContext*>(m_ModelContextPtr.get());
339  if (modelOptions)
340  {
341  isFastMathEnabled = modelOptions->IsFastMathEnabled();
342  }
343  }
344  }
345 #endif
346 
348  reasonIfUnsupported,
349  input,
350  output,
351  descriptor,
352  weights,
353  biases,
354  isFastMathEnabled,
355  nullptr);
356 }
357 
359  const TensorInfo& output,
360  const DepthToSpaceDescriptor& descriptor,
361  Optional<std::string&> reasonIfUnsupported) const
362 {
364  reasonIfUnsupported,
365  input,
366  output,
367  descriptor);
368 }
369 
371  const TensorInfo& output,
372  const DepthwiseConvolution2dDescriptor& descriptor,
373  const TensorInfo& weights,
374  const Optional<TensorInfo>& biases,
375  Optional<std::string&> reasonIfUnsupported) const
376 {
378  reasonIfUnsupported,
379  input,
380  output,
381  descriptor,
382  weights,
383  biases,
384  nullptr);
385 }
386 
388  const TensorInfo& output,
389  Optional<std::string&> reasonIfUnsupported) const
390 {
392  reasonIfUnsupported,
393  input,
394  output);
395 }
396 
398  const TensorInfo& output,
399  const DepthwiseConvolution2dDescriptor& descriptor,
400  const TensorInfo& weights,
401  const Optional<TensorInfo>& biases,
402  Optional<std::string&> reasonIfUnsupported) const
403 {
405  reasonIfUnsupported,
406  input,
407  output,
408  descriptor,
409  weights,
410  biases,
411  nullptr);
412 }
413 
415  const TensorInfo& output,
416  const ElementwiseUnaryDescriptor& descriptor,
417  Optional<std::string&> reasonIfUnsupported) const
418 {
419  switch(descriptor.m_Operation)
420  {
421  case UnaryOperation::Abs:
423  reasonIfUnsupported,
424  input,
425  output);
426  case UnaryOperation::Exp:
428  reasonIfUnsupported,
429  input,
430  output);
431  case UnaryOperation::Neg:
433  reasonIfUnsupported,
434  input,
435  output);
438  reasonIfUnsupported,
439  input,
440  output);
443  reasonIfUnsupported,
444  input,
445  output);
446  default:
447  return false;
448  }
449 }
450 
452  const TensorInfo& output,
453  const FillDescriptor& descriptor,
454  Optional<std::string&> reasonIfUnsupported) const
455 {
456  armnn::IgnoreUnused(input);
457  armnn::IgnoreUnused(output);
458  armnn::IgnoreUnused(descriptor);
459 
460  return IsNeonBackendSupported(reasonIfUnsupported);
461 }
462 
464  const TensorInfo& output,
465  Optional<std::string&> reasonIfUnsupported) const
466 {
467  armnn::IgnoreUnused(output);
468  return IsNeonBackendSupported(reasonIfUnsupported) &&
469  IsSupportedForDataTypeGeneric(reasonIfUnsupported,
470  input.GetDataType(),
471  &FalseFuncF16<>,
472  &TrueFunc<>,
473  &FalseFuncU8<>,
474  &FalseFuncI32<>,
475  &FalseFuncU8<>);
476 }
477 
479  const TensorInfo& output,
480  const TensorInfo& weights,
481  const TensorInfo& biases,
482  const FullyConnectedDescriptor& descriptor,
483  Optional<std::string&> reasonIfUnsupported) const
484 {
486  reasonIfUnsupported,
487  input,
488  output,
489  weights,
490  biases,
491  descriptor,
492  nullptr);
493 }
494 
496  const TensorInfo& input1,
497  const TensorInfo& output,
498  const GatherDescriptor& descriptor,
499  Optional<std::string&> reasonIfUnsupported) const
500 {
502  reasonIfUnsupported,
503  input0,
504  input1,
505  output,
506  descriptor);
507 }
508 
510  const armnn::TensorInfo& input1,
511  const armnn::TensorInfo& output,
512  armnn::Optional<std::string&> reasonIfUnsupported) const
513 {
515  return IsComparisonSupported(input0, input1, output, descriptor, reasonIfUnsupported);
516 }
517 
519  Optional<std::string&> reasonIfUnsupported) const
520 {
521  return IsNeonBackendSupported(reasonIfUnsupported, input);
522 }
523 
525  const TensorInfo& output,
526  const InstanceNormalizationDescriptor& descriptor,
527  Optional<std::string&> reasonIfUnsupported) const
528 {
530  reasonIfUnsupported,
531  input,
532  output,
533  descriptor);
534 }
535 
537  const TensorInfo& output,
538  const L2NormalizationDescriptor& descriptor,
539  Optional<std::string&> reasonIfUnsupported) const
540 {
541  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonL2NormalizationWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
542 }
543 
545  const TensorInfo& input1,
546  const TensorInfo& output,
547  const LogicalBinaryDescriptor& descriptor,
548  Optional<std::string&> reasonIfUnsupported) const
549 {
550  switch(descriptor.m_Operation)
551  {
554  reasonIfUnsupported,
555  input0,
556  input1,
557  output);
560  reasonIfUnsupported,
561  input0,
562  input1,
563  output);
564  default:
565  return false;
566  }
567 }
568 
570  const TensorInfo& output,
571  const LogSoftmaxDescriptor& descriptor,
572  Optional<std::string&> reasonIfUnsupported) const
573 {
574  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonLogSoftmaxWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
575 }
576 
578  const TensorInfo& outputStateIn,
579  const TensorInfo& cellStateIn,
580  const TensorInfo& scratchBuffer,
581  const TensorInfo& outputStateOut,
582  const TensorInfo& cellStateOut,
583  const TensorInfo& output,
584  const LstmDescriptor& descriptor,
585  const LstmInputParamsInfo& paramsInfo,
586  Optional<std::string&> reasonIfUnsupported) const
587 {
589  reasonIfUnsupported,
590  input,
591  outputStateIn,
592  cellStateIn,
593  scratchBuffer,
594  outputStateOut,
595  cellStateOut,
596  output,
597  descriptor,
598  paramsInfo);
599 }
600 
602  const TensorInfo& input1,
603  const TensorInfo& output,
604  Optional<std::string&> reasonIfUnsupported) const
605 {
607  reasonIfUnsupported,
608  input0,
609  input1,
610  output);
611 }
612 
614  const TensorInfo& output,
615  const MeanDescriptor& descriptor,
616  Optional<std::string&> reasonIfUnsupported) const
617 {
619  reasonIfUnsupported,
620  input,
621  output,
622  descriptor);
623 }
624 
625 bool NeonLayerSupport::IsMergerSupported(const std::vector<const TensorInfo*> inputs,
626  const TensorInfo& output,
627  const MergerDescriptor& descriptor,
628  Optional<std::string&> reasonIfUnsupported) const
629 {
630  return IsConcatSupported(inputs, output, descriptor, reasonIfUnsupported);
631 }
632 
634  const TensorInfo& input1,
635  const TensorInfo& output,
636  Optional<std::string&> reasonIfUnsupported) const
637 {
639  reasonIfUnsupported,
640  input0,
641  input1,
642  output);
643 }
644 
646  const TensorInfo& input1,
647  const TensorInfo& output,
648  Optional<std::string&> reasonIfUnsupported) const
649 {
651  reasonIfUnsupported,
652  input0,
653  input1,
654  output,
655  nullptr);
656 }
657 
659  const TensorInfo& input1,
660  const TensorInfo& output,
661  Optional<std::string&> reasonIfUnsupported) const
662 {
664  reasonIfUnsupported,
665  input0,
666  input1,
667  output,
668  nullptr);
669 }
670 
672  const TensorInfo& output,
673  const NormalizationDescriptor& descriptor,
674  Optional<std::string&> reasonIfUnsupported) const
675 {
677  reasonIfUnsupported,
678  input,
679  output,
680  descriptor);
681 }
682 
684  Optional<std::string&> reasonIfUnsupported) const
685 {
686  return IsNeonBackendSupported(reasonIfUnsupported, output);
687 }
688 
690  const TensorInfo& output,
691  const PadDescriptor& descriptor,
692  Optional<std::string&> reasonIfUnsupported) const
693 {
695  reasonIfUnsupported,
696  input,
697  output,
698  descriptor);
699 }
700 
702  const TensorInfo& output,
703  const PermuteDescriptor& descriptor,
704  Optional<std::string&> reasonIfUnsupported) const
705 {
706  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPermuteWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
707 }
708 
710  const TensorInfo& output,
711  const Pooling2dDescriptor& descriptor,
712  Optional<std::string&> reasonIfUnsupported) const
713 {
714  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPooling2dWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
715 }
716 
718  const armnn::TensorInfo &alpha,
719  const armnn::TensorInfo &output,
720  armnn::Optional<std::string &> reasonIfUnsupported) const
721 {
722  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonPreluWorkloadValidate, reasonIfUnsupported, input, alpha, output);
723 }
724 
726  const TensorInfo& previousOutputIn,
727  const TensorInfo& previousCellStateIn,
728  const TensorInfo& outputStateOut,
729  const TensorInfo& cellStateOut,
730  const TensorInfo& output,
731  const QLstmDescriptor& descriptor,
732  const LstmInputParamsInfo& paramsInfo,
733  Optional<std::string&> reasonIfUnsupported) const
734 {
735  // Check required here in order to pass IsLayerSupported for datatypes tests
736  if (input.GetDataType() == armnn::DataType::QAsymmS8 &&
737  previousOutputIn.GetDataType() == armnn::DataType::QAsymmS8 &&
738  previousCellStateIn.GetDataType() == armnn::DataType::QSymmS16 &&
739  outputStateOut.GetDataType() == armnn::DataType::QAsymmS8 &&
740  cellStateOut.GetDataType() == armnn::DataType::QSymmS16 &&
742  {
744  reasonIfUnsupported,
745  input,
746  previousCellStateIn,
747  previousOutputIn,
748  cellStateOut,
749  outputStateOut,
750  output,
751  descriptor,
752  paramsInfo);
753  }
754  else
755  {
756  return false;
757  }
758 }
759 
761  const TensorInfo& output,
762  Optional<std::string&> reasonIfUnsupported) const
763 {
765  reasonIfUnsupported,
766  input,
767  output);
768 }
769 
771  const TensorInfo& cellStateIn,
772  const TensorInfo& outputStateIn,
773  const TensorInfo& cellStateOut,
774  const TensorInfo& outputStateOut,
775  const QuantizedLstmInputParamsInfo& paramsInfo,
776  Optional<std::string&> reasonIfUnsupported) const
777 {
779  reasonIfUnsupported,
780  input,
781  cellStateIn,
782  outputStateIn,
783  cellStateOut,
784  outputStateOut,
785  paramsInfo);
786 }
787 
789  const TensorInfo& output,
790  const ReduceDescriptor& descriptor,
791  Optional<std::string&> reasonIfUnsupported) const
792 {
794  reasonIfUnsupported,
795  input,
796  output,
797  descriptor);
798 }
799 
801  const TensorInfo& output,
802  const ReshapeDescriptor& descriptor,
803  Optional<std::string&> reasonIfUnsupported) const
804 {
805  armnn::IgnoreUnused(descriptor);
807  reasonIfUnsupported,
808  input,
809  output);
810 }
811 
813  const TensorInfo& output,
814  const ResizeDescriptor& descriptor,
815  Optional<std::string&> reasonIfUnsupported) const
816 {
818  reasonIfUnsupported,
819  input,
820  output,
821  descriptor);
822 }
823 
825  const TensorInfo& output,
826  Optional<std::string&> reasonIfUnsupported) const
827 {
828  ResizeDescriptor descriptor;
829  descriptor.m_Method = ResizeMethod::Bilinear;
830  descriptor.m_DataLayout = DataLayout::NCHW;
831 
832  const TensorShape& outputShape = output.GetShape();
833  descriptor.m_TargetHeight = outputShape[2];
834  descriptor.m_TargetWidth = outputShape[3];
835 
836  return IsResizeSupported(input, output, descriptor, reasonIfUnsupported);
837 }
838 
840  const TensorInfo& output,
841  Optional<std::string&> reasonIfUnsupported) const
842 {
844  return IsElementwiseUnarySupported(input, output, descriptor, reasonIfUnsupported);
845 }
846 
848  const TensorInfo& output,
849  const SliceDescriptor& descriptor,
850  Optional<std::string&> reasonIfUnsupported) const
851 {
853  reasonIfUnsupported,
854  input,
855  output,
856  descriptor);
857 }
858 
860  const TensorInfo& output,
861  const SoftmaxDescriptor& descriptor,
862  Optional<std::string&> reasonIfUnsupported) const
863 {
864  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonSoftmaxWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
865 }
866 
868  const TensorInfo& output,
869  const SpaceToBatchNdDescriptor& descriptor,
870  Optional<std::string&> reasonIfUnsupported) const
871 {
873  reasonIfUnsupported,
874  input,
875  output,
876  descriptor);
877 }
878 
880  const TensorInfo& output,
881  const SpaceToDepthDescriptor& descriptor,
882  Optional<std::string&> reasonIfUnsupported) const
883 {
885  reasonIfUnsupported,
886  input,
887  output,
888  descriptor);
889 }
890 
892  const ViewsDescriptor& descriptor,
893  Optional<std::string&> reasonIfUnsupported) const
894 {
895  armnn::IgnoreUnused(descriptor);
896  return IsSupportedForDataTypeNeon(reasonIfUnsupported,
897  input.GetDataType(),
898  &TrueFunc<>,
899  &TrueFunc<>);
900 }
901 
903  const std::vector<std::reference_wrapper<TensorInfo>>& outputs,
904  const ViewsDescriptor& descriptor,
905  Optional<std::string&> reasonIfUnsupported) const
906 {
907 #if defined(ARMCOMPUTENEON_ENABLED)
908  // Split along the last dimension, cannot use sub-tensors
909  // as width and height of the sub-tensors do not match
910  // the width and height of the parent tensor
911  // in case of input with more than 2D.
912  std::set<unsigned int> splitAxis = ComputeSplitAxis(descriptor, input.GetShape());
913  if (descriptor.GetNumDimensions() > 2 && splitAxis.size() == 1 &&
914  *splitAxis.begin() == descriptor.GetNumDimensions() - 1 )
915  {
917  reasonIfUnsupported,
918  input,
919  outputs,
920  *splitAxis.begin());
921  }
922 #endif
923  IgnoreUnused(descriptor);
924  for (auto output : outputs)
925  {
926  if (!input.IsTypeSpaceMatch(output)) // Cannot use sub-tensors if the types are not same space
927  {
928  SetValueChecked(reasonIfUnsupported, "Neon Splitter: Types and quantization parameters must match.");
929  return false;
930  }
931  }
932  return true;
933 }
934 
935 bool NeonLayerSupport::IsStackSupported(const std::vector<const TensorInfo*>& inputs,
936  const TensorInfo& output,
937  const StackDescriptor& descriptor,
938  Optional<std::string&> reasonIfUnsupported) const
939 {
941  reasonIfUnsupported,
942  inputs,
943  output,
944  descriptor);
945 }
946 
948  const TensorInfo& output,
949  const StridedSliceDescriptor& descriptor,
950  Optional<std::string&> reasonIfUnsupported) const
951 {
953  reasonIfUnsupported,
954  input,
955  output,
956  descriptor);
957 }
958 
960  const TensorInfo& input1,
961  const TensorInfo& output,
962  Optional<std::string&> reasonIfUnsupported) const
963 {
965  reasonIfUnsupported,
966  input0,
967  input1,
968  output,
969  nullptr);
970 }
971 
973  const TensorInfo& output,
974  const TransposeConvolution2dDescriptor& descriptor,
975  const TensorInfo& weights,
976  const Optional<TensorInfo>& biases,
977  Optional<std::string&> reasonIfUnsupported) const
978 {
980  reasonIfUnsupported,
981  input,
982  output,
983  descriptor,
984  weights,
985  biases);
986 }
987 
989  const TensorInfo& output,
990  const TransposeDescriptor& descriptor,
991  Optional<std::string&> reasonIfUnsupported) const
992 {
993  FORWARD_WORKLOAD_VALIDATE_FUNC(NeonTransposeWorkloadValidate, reasonIfUnsupported, input, output, descriptor);
994 }
995 
996 } // namespace armnn
arm_compute::Status NeonGatherWorkloadValidate(const TensorInfo &input, const TensorInfo &indices, const TensorInfo &output, const GatherDescriptor &descriptor)
arm_compute::Status NeonNegWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
UnaryOperation m_Operation
Specifies the elementwiseUnary operation to execute.
bool IsSliceSupported(const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsSoftmaxSupported(const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsConvolution2dSupported(const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A ViewsDescriptor for the SplitterLayer.
bool IsTypeSpaceMatch(const TensorInfo &other) const
Check that the types are the same and, if quantize, that the quantization parameters are the same...
Definition: Tensor.cpp:423
bool IsConvertFp32ToFp16Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonSpaceToDepthWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor)
arm_compute::Status NeonSplitterWorkloadValidate(const TensorInfo &input, const std::vector< std::reference_wrapper< TensorInfo >> &outputs, unsigned int splitAxis)
A TransposeConvolution2dDescriptor for the TransposeConvolution2dLayer.
const TensorShape & GetShape() const
Definition: Tensor.hpp:187
arm_compute::Status NeonLogSoftmaxWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor)
A ReshapeDescriptor for the ReshapeLayer.
bool IsGatherSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const GatherDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported) const override
arm_compute::Status NeonBatchNormalizationValidate(const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, const ActivationDescriptor *activationDescriptor)
bool IsDilatedDepthwiseConvolutionSupported(const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string &> reason=EmptyOptional()) const override
bool IsQuantizedLstmSupported(const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const QuantizedLstmInputParamsInfo &paramsInfo, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
uint32_t GetNumDimensions() const
Get the number of dimensions.
A ComparisonDescriptor for the ComparisonLayer.
Definition: Descriptors.hpp:78
bool IsDepthToSpaceSupported(const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsL2NormalizationSupported(const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsStridedSliceSupported(const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsConvertFp32ToBf16Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
ISubgraphViewConverter supported
bool IsAdditionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A Convolution2dDescriptor for the Convolution2dLayer.
bool IsConcatSupported(const std::vector< const TensorInfo *> inputs, const TensorInfo &output, const ConcatDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsSplitterSupported(const TensorInfo &input, const ViewsDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonDepthwiseConvolutionWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, const ActivationDescriptor *activationDescriptor)
arm_compute::Status NeonMeanWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &desc)
ResizeMethod m_Method
The Interpolation method to use (Bilinear, NearestNeighbor).
arm_compute::Status NeonActivationWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor)
arm_compute::Status NeonMinimumWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output)
Validate function for validating the inputs and output.
arm_compute::Status NeonStridedSliceWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const StridedSliceDescriptor &descriptor)
arm_compute::Status NeonTransposeConvolution2dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases)
bool IsGreaterSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonLstmFloatWorkloadValidate(const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo)
arm_compute::Status NeonQLstmWorkloadValidate(const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo)
bool IsResizeBilinearSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonSliceWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const SliceDescriptor &descriptor)
A LogicalBinaryDescriptor for the LogicalBinaryLayer.
arm_compute::Status NeonFullyConnectedWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const TensorInfo &biases, const FullyConnectedDescriptor &descriptor, const ActivationDescriptor *activationDescriptor)
bool IsConstantSupported(const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
Copyright (c) 2021 ARM Limited and Contributors.
arm_compute::Status NeonQuantizeWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsPadSupported(const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
void IgnoreUnused(Ts &&...)
std::set< unsigned int > ComputeSplitAxis(const armnn::SplitterDescriptor &desc, const TensorShape &input)
arm_compute::Status NeonAdditionWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor)
A SpaceToDepthDescriptor for the SpaceToDepthLayer.
arm_compute::Status NeonLogicalAndWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output)
arm_compute::Status NeonInstanceNormalizationWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor)
bool IsPooling2dSupported(const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonLogicalOrWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output)
LogicalBinaryOperation m_Operation
Specifies the logical operation to execute.
A BatchToSpaceNdDescriptor for the BatchToSpaceNdLayer.
bool IsActivationSupported(const TensorInfo &input, const TensorInfo &output, const ActivationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
The NeonBackendModelContext is used to pass in Neon specific backend ModelOptions.
bool IsMultiplicationSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsComparisonSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A ResizeDescriptor for the ResizeLayer.
arm_compute::Status NeonL2NormalizationWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const L2NormalizationDescriptor &descriptor)
arm_compute::Status NeonAbsWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
A StackDescriptor for the StackLayer.
arm_compute::Status NeonQuantizedLstmWorkloadValidate(const TensorInfo &input, const TensorInfo &cellStateIn, const TensorInfo &outputStateIn, const TensorInfo &cellStateOut, const TensorInfo &outputStateOut, const QuantizedLstmInputParamsInfo &paramsInfo)
arm_compute::Status NeonStackWorkloadValidate(const std::vector< const TensorInfo *> &inputs, const TensorInfo &output, const StackDescriptor &descriptor)
arm_compute::Status NeonSpaceToBatchNdWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor)
arm_compute::Status NeonSubtractionWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor)
A PadDescriptor for the PadLayer.
arm_compute::Status NeonReduceWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &desc)
bool IsAbsSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsLogicalBinarySupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const LogicalBinaryDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported) const override
DataType
Definition: Types.hpp:32
bool IsFillSupported(const TensorInfo &input, const TensorInfo &output, const FillDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
An LstmDescriptor for the LstmLayer.
arm_compute::Status NeonExpWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
arm_compute::Status NeonReshapeWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsInputSupported(const TensorInfo &input, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsStackSupported(const std::vector< const TensorInfo *> &inputs, const TensorInfo &output, const StackDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
std::shared_ptr< IBackendModelContext > IBackendSpecificModelContextPtr
A L2NormalizationDescriptor for the L2NormalizationLayer.
bool IsMinimumSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsPreluSupported(const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsOutputSupported(const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
An ArgMinMaxDescriptor for ArgMinMaxLayer.
Definition: Descriptors.hpp:56
bool IsInstanceNormalizationSupported(const TensorInfo &input, const TensorInfo &output, const InstanceNormalizationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
DataType GetDataType() const
Definition: Tensor.hpp:194
An OriginsDescriptor for the ConcatLayer.
A ReduceDescriptor for the REDUCE operators.
A FullyConnectedDescriptor for the FullyConnectedLayer.
arm_compute::Status NeonLogicalNotWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsMergerSupported(const std::vector< const TensorInfo *> inputs, const TensorInfo &output, const MergerDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsNormalizationSupported(const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsElementwiseUnarySupported(const TensorInfo &input, const TensorInfo &output, const ElementwiseUnaryDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
uint32_t m_TargetWidth
Target width value.
A GatherDescriptor for the GatherLayer.
Status
enumeration
Definition: Types.hpp:26
arm_compute::Status NeonComparisonWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ComparisonDescriptor &descriptor)
bool IsDivisionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonConcatWorkloadValidate(const std::vector< const TensorInfo *> &inputs, const TensorInfo &output, const OriginsDescriptor &descriptor)
arm_compute::Status NeonPermuteWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor)
bool IsFloorSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A QLstmDescriptor for the QLstmLayer.
arm_compute::Status NeonConvolution2dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor)
bool IsLogSoftmaxSupported(const TensorInfo &input, const TensorInfo &output, const LogSoftmaxDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
An ActivationDescriptor for the ActivationLayer.
Definition: Descriptors.hpp:25
bool IsResizeSupported(const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsTransposeConvolution2dSupported(const TensorInfo &input, const TensorInfo &output, const TransposeConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
uint32_t m_TargetHeight
Target height value.
A SliceDescriptor for the SliceLayer.
arm_compute::Status NeonDequantizeWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsReshapeSupported(const TensorInfo &input, const TensorInfo &output, const ReshapeDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonDivisionWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor)
bool IsBatchNormalizationSupported(const TensorInfo &input, const TensorInfo &output, const TensorInfo &mean, const TensorInfo &var, const TensorInfo &beta, const TensorInfo &gamma, const BatchNormalizationDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
void SetValueChecked(Optional< T &> optionalRef, V &&val)
A SpaceToBatchNdDescriptor for the SpaceToBatchNdLayer.
arm_compute::Status NeonNormalizationWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const NormalizationDescriptor &descriptor)
arm_compute::Status NeonBatchToSpaceNdWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &desc)
bool IsQuantizeSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonRsqrtWorkloadValidate(const TensorInfo &input, const TensorInfo &output)
bool IsConvertFp16ToFp32Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonPadWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const PadDescriptor &descriptor)
A ElementwiseUnaryDescriptor for the ElementwiseUnaryLayer.
Definition: Descriptors.hpp:98
bool IsSubtractionSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsLstmSupported(const TensorInfo &input, const TensorInfo &outputStateIn, const TensorInfo &cellStateIn, const TensorInfo &scratchBuffer, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const LstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonArgMinMaxWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor)
arm_compute::Status NeonSoftmaxWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const SoftmaxDescriptor &descriptor)
bool IsSpaceToBatchNdSupported(const TensorInfo &input, const TensorInfo &output, const SpaceToBatchNdDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonTransposeWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor)
uint32_t GetNumDimensions() const
Get the number of dimensions.
A MeanDescriptor for the MeanLayer.
bool IsDepthwiseConvolutionSupported(const TensorInfo &input, const TensorInfo &output, const DepthwiseConvolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonMaximumWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output)
bool IsArgMinMaxSupported(const TensorInfo &input, const TensorInfo &output, const ArgMinMaxDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsSpaceToDepthSupported(const TensorInfo &input, const TensorInfo &output, const SpaceToDepthDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A TransposeDescriptor for the TransposeLayer.
A StridedSliceDescriptor for the StridedSliceLayer.
bool IsTransposeSupported(const TensorInfo &input, const TensorInfo &output, const TransposeDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsMaximumSupported(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonConstantWorkloadValidate(const TensorInfo &output)
bool IsFullyConnectedSupported(const TensorInfo &input, const TensorInfo &output, const TensorInfo &weights, const TensorInfo &biases, const FullyConnectedDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsMeanSupported(const TensorInfo &input, const TensorInfo &output, const MeanDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsBatchToSpaceNdSupported(const TensorInfo &input, const TensorInfo &output, const BatchToSpaceNdDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonPreluWorkloadValidate(const TensorInfo &input, const TensorInfo &alpha, const TensorInfo &output)
#define FORWARD_WORKLOAD_VALIDATE_FUNC(func, reasonIfUnsupported,...)
arm_compute::Status NeonDepthToSpaceWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const DepthToSpaceDescriptor &descriptor)
A Pooling2dDescriptor for the Pooling2dLayer.
arm_compute::Status NeonResizeWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const ResizeDescriptor &descriptor)
A NormalizationDescriptor for the NormalizationLayer.
bool IsQLstmSupported(const TensorInfo &input, const TensorInfo &previousOutputIn, const TensorInfo &previousCellStateIn, const TensorInfo &outputStateOut, const TensorInfo &cellStateOut, const TensorInfo &output, const QLstmDescriptor &descriptor, const LstmInputParamsInfo &paramsInfo, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
bool IsConvertBf16ToFp32Supported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
DataLayout m_DataLayout
The data layout to be used (NCHW, NHWC).
An InstanceNormalizationDescriptor for InstanceNormalizationLayer.
unsigned int GetConcatAxis() const
Get the concatenation axis value.
bool IsSupportedForDataTypeGeneric(Optional< std::string &> reasonIfUnsupported, DataType dataType, Float16Func float16FuncPtr, Float32Func float32FuncPtr, Uint8Func uint8FuncPtr, Int32Func int32FuncPtr, BooleanFunc booleanFuncPtr, Params &&... params)
A SoftmaxDescriptor for the SoftmaxLayer.
bool IsReduceSupported(const TensorInfo &input, const TensorInfo &output, const ReduceDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
arm_compute::Status NeonPooling2dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const Pooling2dDescriptor &descriptor)
bool IsDequantizeSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A DepthwiseConvolution2dDescriptor for the DepthwiseConvolution2dLayer.
bool IsRsqrtSupported(const TensorInfo &input, const TensorInfo &output, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override
A FillDescriptor for the FillLayer.
A BatchNormalizationDescriptor for the BatchNormalizationLayer.
arm_compute::Status NeonMultiplicationWorkloadValidate(const TensorInfo &input0, const TensorInfo &input1, const TensorInfo &output, const ActivationDescriptor *activationDescriptor)
A PermuteDescriptor for the PermuteLayer.
bool IsPermuteSupported(const TensorInfo &input, const TensorInfo &output, const PermuteDescriptor &descriptor, Optional< std::string &> reasonIfUnsupported=EmptyOptional()) const override