23.05
|
Go to the documentation of this file.
16 #include <type_traits>
27 default:
return "Unknown";
47 default:
return "Unknown";
57 default:
return "Unknown";
71 default:
return "Unknown";
85 default:
return "Unknown";
101 default:
return "Unknown";
111 default:
return "Unknown";
122 default:
return "Unknown";
132 default:
return "Unknown";
142 default:
return "Unknown";
153 default:
return "Unknown";
159 switch (reduce_operation)
166 default:
return "Unknown";
187 template <
unsigned N>
188 constexpr
bool StrEqual(
const char* strA,
const char (&strB)[N])
191 for (
unsigned i = 0; isEqual && (i < N); ++i)
193 isEqual = (strA[i] == strB[i]);
248 default:
return "Unknown";
258 default:
return "Unknown";
268 default:
return "Unknown";
278 default:
return "Unknown";
284 switch (memBlockStrategyType)
288 default:
return "Unknown";
294 : std::integral_constant<bool, std::is_floating_point<T>::value && sizeof(T) == 2>
300 return std::is_integral<T>::value;
345 os <<
"Dimensionality Not Specified";
357 template<
typename QuantizedType>
358 QuantizedType
Quantize(
float value,
float scale, int32_t offset);
366 template <
typename QuantizedType>
367 float Dequantize(QuantizedType value,
float scale, int32_t offset);
371 if (
info.GetDataType() != dataType)
373 std::stringstream ss;
375 <<
" for tensor:" <<
info.GetShape()
NormalizationAlgorithmMethod
constexpr bool IsQuantized8BitType(DataType dataType)
constexpr char const * GetArgMinMaxFunctionAsCString(ArgMinMaxFunction function)
NormalizationAlgorithmChannel
Dimensionality GetDimensionality() const
Function that returns the tensor type.
Base class for all ArmNN exceptions so that users can filter to just those.
constexpr char const * GetOutputShapeRoundingAsCString(OutputShapeRounding rounding)
constexpr char const * GetPoolingAlgorithmAsCString(PoolingAlgorithm pooling)
@ LocalContrast
Jarret 2009: Local Contrast Normalization.
constexpr bool IsQuantizedType()
@ CpuAcc
CPU Execution: NEON: ArmCompute.
@ GpuAcc
GPU Execution: OpenCL: ArmCompute.
constexpr char const * GetUnaryOperationAsCString(UnaryOperation operation)
@ IgnoreValue
The padding fields count, but are ignored.
Compute
The Compute enum is now deprecated and it is now being replaced by BackendId.
constexpr unsigned int GetDataTypeSize(DataType dataType)
Copyright (c) 2021 ARM Limited and Contributors.
constexpr char const * GetReduceOperationAsCString(ReduceOperation reduce_operation)
constexpr char const * GetLogicalBinaryOperationAsCString(LogicalBinaryOperation operation)
constexpr armnn::Compute ParseComputeDevice(const char *str)
Deprecated function that will be removed together with the Compute enum.
constexpr const char * GetResizeMethodAsCString(ResizeMethod method)
@ LocalBrightness
Krichevsky 2012: Local Brightness Normalization.
bool GetDimensionSpecificity(unsigned int i) const
Gets information about if the dimension size has been specified or not.
constexpr char const * GetPaddingModeAsCString(PaddingMode mode)
PaddingMode
The padding mode controls whether the padding should be filled with constant values (Constant),...
constexpr bool StrEqual(const char *strA, const char(&strB)[N])
constexpr char const * GetComparisonOperationAsCString(ComparisonOperation operation)
PaddingMethod
The padding method modifies the output of pooling layers.
constexpr const char * GetNormalizationAlgorithmChannelAsCString(NormalizationAlgorithmChannel channel)
void VerifyTensorInfoDataType(const armnn::TensorInfo &info, armnn::DataType dataType)
constexpr char const * GetActivationFunctionAsCString(ActivationFunction activation)
constexpr char const * GetStatusAsCString(Status status)
unsigned int GetNumDimensions() const
Function that returns the tensor rank.
constexpr const char * GetNormalizationAlgorithmMethodAsCString(NormalizationAlgorithmMethod method)
QuantizedType Quantize(float value, float scale, int32_t offset)
Quantize a floating point data type into an 8-bit data type.
constexpr char const * GetPaddingMethodAsCString(PaddingMethod method)
std::ostream & operator<<(std::ostream &os, const std::vector< Compute > &compute)
Deprecated function that will be removed together with the Compute enum.
@ Exclude
The padding fields don't count and are ignored.
float Dequantize(QuantizedType value, float scale, int32_t offset)
Dequantize an 8-bit data type into a floating point data type.
constexpr const char * GetDataTypeName(DataType dataType)
constexpr const char * GetDataLayoutName(DataLayout dataLayout)
@ BoundedReLu
min(a, max(b, input)) ReLu1 & ReLu6.
@ CpuRef
CPU Execution: Reference C++ kernels.
constexpr char const * GetBinaryOperationAsCString(BinaryOperation operation)
constexpr const char * GetMemBlockStrategyTypeName(MemBlockStrategyType memBlockStrategyType)