23.02
|
Go to the documentation of this file.
16 #include <type_traits>
27 default:
return "Unknown";
47 default:
return "Unknown";
57 default:
return "Unknown";
71 default:
return "Unknown";
87 default:
return "Unknown";
97 default:
return "Unknown";
108 default:
return "Unknown";
118 default:
return "Unknown";
128 default:
return "Unknown";
139 default:
return "Unknown";
145 switch (reduce_operation)
152 default:
return "Unknown";
173 template <
unsigned N>
174 constexpr
bool StrEqual(
const char* strA,
const char (&strB)[N])
177 for (
unsigned i = 0; isEqual && (i < N); ++i)
179 isEqual = (strA[i] == strB[i]);
234 default:
return "Unknown";
244 default:
return "Unknown";
254 default:
return "Unknown";
264 default:
return "Unknown";
270 switch (memBlockStrategyType)
274 default:
return "Unknown";
280 : std::integral_constant<bool, std::is_floating_point<T>::value && sizeof(T) == 2>
286 return std::is_integral<T>::value;
331 os <<
"Dimensionality Not Specified";
343 template<
typename QuantizedType>
344 QuantizedType
Quantize(
float value,
float scale, int32_t offset);
352 template <
typename QuantizedType>
353 float Dequantize(QuantizedType value,
float scale, int32_t offset);
357 if (
info.GetDataType() != dataType)
359 std::stringstream ss;
361 <<
" for tensor:" <<
info.GetShape()
NormalizationAlgorithmMethod
constexpr bool IsQuantized8BitType(DataType dataType)
constexpr char const * GetArgMinMaxFunctionAsCString(ArgMinMaxFunction function)
NormalizationAlgorithmChannel
Dimensionality GetDimensionality() const
Function that returns the tensor type.
Base class for all ArmNN exceptions so that users can filter to just those.
constexpr char const * GetOutputShapeRoundingAsCString(OutputShapeRounding rounding)
constexpr char const * GetPoolingAlgorithmAsCString(PoolingAlgorithm pooling)
@ LocalContrast
Jarret 2009: Local Contrast Normalization.
constexpr bool IsQuantizedType()
@ CpuAcc
CPU Execution: NEON: ArmCompute.
@ GpuAcc
GPU Execution: OpenCL: ArmCompute.
constexpr char const * GetUnaryOperationAsCString(UnaryOperation operation)
@ IgnoreValue
The padding fields count, but are ignored.
Compute
The Compute enum is now deprecated and it is now being replaced by BackendId.
constexpr unsigned int GetDataTypeSize(DataType dataType)
Copyright (c) 2021 ARM Limited and Contributors.
constexpr char const * GetReduceOperationAsCString(ReduceOperation reduce_operation)
constexpr char const * GetLogicalBinaryOperationAsCString(LogicalBinaryOperation operation)
constexpr armnn::Compute ParseComputeDevice(const char *str)
Deprecated function that will be removed together with the Compute enum.
constexpr const char * GetResizeMethodAsCString(ResizeMethod method)
@ LocalBrightness
Krichevsky 2012: Local Brightness Normalization.
bool GetDimensionSpecificity(unsigned int i) const
Gets information about if the dimension size has been specified or not.
constexpr char const * GetPaddingModeAsCString(PaddingMode mode)
PaddingMode
The padding mode controls whether the padding should be filled with constant values (Constant),...
constexpr bool StrEqual(const char *strA, const char(&strB)[N])
constexpr char const * GetComparisonOperationAsCString(ComparisonOperation operation)
PaddingMethod
The padding method modifies the output of pooling layers.
constexpr const char * GetNormalizationAlgorithmChannelAsCString(NormalizationAlgorithmChannel channel)
void VerifyTensorInfoDataType(const armnn::TensorInfo &info, armnn::DataType dataType)
constexpr char const * GetActivationFunctionAsCString(ActivationFunction activation)
constexpr char const * GetStatusAsCString(Status status)
unsigned int GetNumDimensions() const
Function that returns the tensor rank.
constexpr const char * GetNormalizationAlgorithmMethodAsCString(NormalizationAlgorithmMethod method)
QuantizedType Quantize(float value, float scale, int32_t offset)
Quantize a floating point data type into an 8-bit data type.
constexpr char const * GetPaddingMethodAsCString(PaddingMethod method)
std::ostream & operator<<(std::ostream &os, const std::vector< Compute > &compute)
Deprecated function that will be removed together with the Compute enum.
@ Exclude
The padding fields don't count and are ignored.
float Dequantize(QuantizedType value, float scale, int32_t offset)
Dequantize an 8-bit data type into a floating point data type.
constexpr const char * GetDataTypeName(DataType dataType)
constexpr const char * GetDataLayoutName(DataLayout dataLayout)
@ BoundedReLu
min(a, max(b, input)) ReLu1 & ReLu6.
@ CpuRef
CPU Execution: Reference C++ kernels.
constexpr const char * GetMemBlockStrategyTypeName(MemBlockStrategyType memBlockStrategyType)