diff options
author | Sang-Hoon Park <sang-hoon.park@arm.com> | 2021-05-05 10:34:47 +0100 |
---|---|---|
committer | Sheri Zhang <sheri.zhang@arm.com> | 2021-05-05 11:43:47 +0000 |
commit | c9309f22a026dfce92365e2f0802c40e8e1c449e (patch) | |
tree | c3d1b8498bb319d8716dd78979b24fd24ae8c411 /docs/09_operators_list.dox | |
parent | d813bab10bb4fe954fa0e962e1402ed1377617da (diff) | |
download | ComputeLibrary-c9309f22a026dfce92365e2f0802c40e8e1c449e.tar.gz |
Restructure Documentation (Part 2)
The followings are done:
- Move operator list documnetation
- Introduction page is moved to the top
- The sections for experimental API and programming
model are merged into library architecture page.
Resolves: COMPMID-4198
Change-Id: Iad824d6c8ba8d31e0bf76afd3fb67abbe32a1667
Signed-off-by: Sang-Hoon Park <sang-hoon.park@arm.com>
Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/5570
Tested-by: Arm Jenkins <bsgcomp@arm.com>
Reviewed-by: Michele Di Giorgio <michele.digiorgio@arm.com>
Reviewed-by: Sheri Zhang <sheri.zhang@arm.com>
Comments-Addressed: Arm Jenkins <bsgcomp@arm.com>
Diffstat (limited to 'docs/09_operators_list.dox')
-rw-r--r-- | docs/09_operators_list.dox | 3123 |
1 files changed, 0 insertions, 3123 deletions
diff --git a/docs/09_operators_list.dox b/docs/09_operators_list.dox deleted file mode 100644 index fc41265738..0000000000 --- a/docs/09_operators_list.dox +++ /dev/null @@ -1,3123 +0,0 @@ -/// -/// Copyright (c) 2021 Arm Limited. -/// -/// SPDX-License-Identifier: MIT -/// -/// Permission is hereby granted, free of charge, to any person obtaining a copy -/// of this software and associated documentation files (the "Software"), to -/// deal in the Software without restriction, including without limitation the -/// rights to use, copy, modify, merge, publish, distribute, sublicense, and/or -/// sell copies of the Software, and to permit persons to whom the Software is -/// furnished to do so, subject to the following conditions: -/// -/// The above copyright notice and this permission notice shall be included in all -/// copies or substantial portions of the Software. -/// -/// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR -/// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, -/// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE -/// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER -/// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, -/// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE -/// SOFTWARE. -/// -namespace arm_compute -{ -/** -@page operators_list Supported Operators - -@tableofcontents - -@section S9_1_operators_list Supported Operators - -Compute Library supports operators that are listed in below table. - -Compute Library supports a wide list of data-types, information can been directly found in the documentation of each kernel/function. -The main data-types that the Machine Learning functions support are the following: - <ul> - <li>BFLOAT16: 16-bit non-standard brain floating point - <li>QASYMM8: 8-bit unsigned asymmetric quantized - <li>QASYMM8_SIGNED: 8-bit signed asymmetric quantized - <li>QSYMM8_PER_CHANNEL: 8-bit signed symmetric quantized (Used for the weights) - <li>QSYMM8: 8-bit unsigned symmetric quantized - <li>QSYMM16: 16-bit unsigned symmetric quantized - <li>F32: 32-bit single precision floating point - <li>F16: 16-bit half precision floating point - <li>S32: 32-bit signed integer - <li>U8: 8-bit unsigned char - <li>All: include all above data types - </ul> - -Compute Library supports the following data layouts (fast changing dimension from right to left): - <ul> - <li>NHWC: The native layout of Compute Library that delivers the best performance where channels are in the fastest changing dimension - <li>NCHW: Legacy layout where width is in the fastest changing dimension - <li>All: include all above data layouts - </ul> -where N = batches, C = channels, H = height, W = width - -<table> -<caption id="multi_row"></caption> -<tr> - <th>Function - <th>Description - <th>Equivalent Android NNAPI Op - <th>Backends - <th>Data Layouts - <th>Data Types -<tr> - <td rowspan="2">ActivationLayer - <td rowspan="2" style="width:200px;"> Function to simulate an activation layer with the specified activation function. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_ELU - <li>ANEURALNETWORKS_HARD_SWISH - <li>ANEURALNETWORKS_LOGISTIC - <li>ANEURALNETWORKS_RELU - <li>ANEURALNETWORKS_RELU1 - <li>ANEURALNETWORKS_RELU6 - <li>ANEURALNETWORKS_TANH - </ul> - <td>NEActivationLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16 - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLActivationLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16 - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">ArgMinMaxLayer - <td rowspan="2" style="width:200px;"> Function to calculate the index of the minimum or maximum values in a tensor based on an axis. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_ARGMAX - <li>ANEURALNETWORKS_ARGMIN - </ul> - <td>NEArgMinMaxLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>U32, S32 - <tr><td>QASYMM8_SIGNED<td>U32, S32 - <tr><td>S32<td>U32, S32 - <tr><td>F16<td>U32, S32 - <tr><td>F32<td>U32, S32 - </table> -<tr> - <td>CLArgMinMaxLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>U32, S32 - <tr><td>QASYMM8_SIGNED<td>U32, S32 - <tr><td>S32<td>U32, S32 - <tr><td>F16<td>U32, S32 - <tr><td>F32<td>U32, S32 - </table> -<tr> - <td rowspan="1">ArithmeticAddition - <td rowspan="1" style="width:200px;"> Function to add 2 tensors. - <td rowspan="1"> - <ul> - <li>ANEURALNETWORKS_ADD - </ul> - <td>NEArithmeticAddition - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16 - <tr><td>QSYMM16<td>QSYMM16<td>S32 - <tr><td>U8<td>U8<td>U8 - <tr><td>U8<td>U8<td>S16 - <tr><td>U8<td>S16<td>S16 - <tr><td>S16<td>U8<td>S16 - <tr><td>S16<td>S16<td>S16 - <tr><td>S32<td>S32<td>S32 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td rowspan="1">ArithmeticSubtraction - <td rowspan="1" style="width:200px;"> Function to substract 2 tensors. - <td rowspan="1"> - <ul> - <li>ANEURALNETWORKS_SUB - </ul> - <td>NEArithmeticSubtraction - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16 - <tr><td>QSYMM16<td>QSYMM16<td>S32 - <tr><td>U8<td>U8<td>U8 - <tr><td>U8<td>U8<td>S16 - <tr><td>U8<td>S16<td>S16 - <tr><td>S16<td>U8<td>S16 - <tr><td>S16<td>S16<td>S16 - <tr><td>S32<td>S32<td>S32 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td rowspan="2">BatchNormalizationLayer - <td rowspan="2" style="width:200px;"> Function to perform batch normalization. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEBatchNormalizationLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td>CLBatchNormalizationLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td rowspan="2">BatchToSpaceLayer - <td rowspan="2" style="width:200px;"> Batch to space transformation. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_BATCH_TO_SPACE_ND - </ul> - <td>NEBatchToSpaceLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>All<td>s32<td>All - </table> -<tr> - <td>CLBatchToSpaceLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>All<td>s32<td>All - </table> -<tr> - <td rowspan="2">BitwiseAnd - <td rowspan="2" style="width:200px;"> Function to performe bitwise AND between 2 tensors. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_LOGICAL_AND - </ul> - <td>NEBitwiseAnd - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td>CLBitwiseAnd - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td rowspan="2">BitwiseNot - <td rowspan="2" style="width:200px;"> Function to performe bitwise NOT. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_LOGICAL_NOT - </ul> - <td>NEBitwiseNot - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td>CLBitwiseNot - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td rowspan="2">BitwiseOr - <td rowspan="2" style="width:200px;"> Function to performe bitwise OR between 2 tensors. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_LOGICAL_OR - </ul> - <td>NEBitwiseOr - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td>CLBitwiseOr - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td rowspan="2">BitwiseXor - <td rowspan="2" style="width:200px;"> Function to performe bitwise XOR between 2 tensors. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEBitwiseXor - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td>CLBitwiseXor - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td rowspan="2">BoundingBoxTransform - <td rowspan="2" style="width:200px;"> Transform proposal bounding boxes to target bounding box using bounding box deltas. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEBoundingBoxTransform - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM16<td>QASYMM8<td>QASYMM16 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLBoundingBoxTransform - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM16<td>QASYMM8<td>QASYMM16 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td rowspan="2">Cast - <td rowspan="2" style="width:200px;"> Function to cast a tensor. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_CAST - </ul> - <td>NECast - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8_SIGNED<td>S16, S32, F32, F16 - <tr><td>QASYMM8<td>U16, S16, S32, F32, F16 - <tr><td>U8<td>U16, S16, S32, F32, F16 - <tr><td>U16<td>U8, U32 - <tr><td>S16<td>QASYMM8_SIGNED, U8, S32 - <tr><td>F16<td>QASYMM8_SIGNED, QASYMM8, F32, S32, U8 - <tr><td>S32<td>QASYMM8_SIGNED, QASYMM8, F16, F32, U8 - <tr><td>F32<td>QASYMM8_SIGNED, QASYMM8, BFLOAT16, F16, S32, U8 - </table> -<tr> - <td>CLCast - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>S8, U16, S16, U32, S32, F16, F32 - <tr><td>U16<td>U8, S8, S16, U32, S32, F16, F32 - <tr><td>S16<td>U8, S8, U16, U32, S32, F16, F32 - <tr><td>U32<td>U8, S8, U16, S16, S32, F16, F32 - <tr><td>S32<td>U8, S8, U16, S16, U32, F16, F32 - <tr><td>F16<td>U8, S8, U16, S16, U32, F32 - <tr><td>F32<td>U8, S8, U16, S16, U32, F16 - </table> -<tr> - <td rowspan="2">ChannelShuffleLayer - <td rowspan="2" style="width:200px;"> Function to shuffle the channels of the input tensor. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_CHANNEL_SHUFFLE - </ul> - <td>NEChannelShuffleLayer - <td> - <ul> - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLChannelShuffleLayer - <td> - <ul> - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="1">Comparison - <td rowspan="1" style="width:200px;"> Function to compare 2 tensors. - <td rowspan="1"> - <ul> - <li>ANEURALNETWORKS_EQUAL - <li>ANEURALNETWORKS_GREATER - <li>ANEURALNETWORKS_GREATER_EQUAL - <li>ANEURALNETWORKS_LESS - <li>ANEURALNETWORKS_LESS_EQUAL - <li>ANEURALNETWORKS_NOT_EQUAL - </ul> - <td>CLComparison - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>All<td>All<td>U8 - </table> -<tr> - <td rowspan="2">ConcatenateLayer - <td rowspan="2" style="width:200px;"> Function to concatenate tensors along a given axis. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_CONCATENATION - </ul> - <td>NEConcatenateLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLConcatenateLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">ConvertFullyConnectedWeights - <td rowspan="2" style="width:200px;"> Function to tranpose the wieghts for the fully connected layer. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEConvertFullyConnectedWeights - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLConvertFullyConnectedWeights - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">ConvolutionLayer - <td rowspan="2" style="width:200px;"> Function to compute a convolution layer. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_CONV_2D - </ul> - <td>NEConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td>CLConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="2">Copy - <td rowspan="2" style="width:200px;"> Function to copy a tensor. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NECopy - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLCopy - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="1">Crop - <td rowspan="1" style="width:200px;"> Performs a copy of input tensor to the output tensor. - <td rowspan="1"> - <ul> - <li>n/a - </ul> - <td>CLCrop - <td> - <ul> - <li>NHWC - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>F32 - </table> -<tr> - <td rowspan="2">CropResize - <td rowspan="2" style="width:200px;"> Function to perform cropping and resizing. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NECropResize - <td> - <ul> - <li>NHWC - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>All<td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLCropResize - <td> - <ul> - <li>NHWC - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>All<td>F32<td>F32<td>F32 - </table> -<tr> - <td rowspan="2">DeconvolutionLayer - <td rowspan="2" style="width:200px;"> Function to compute a deconvolution or tranpose convolution. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D - </ul> - <td>NEDeconvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td>CLDeconvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="1">DeconvolutionLayerUpsample - <td rowspan="1" style="width:200px;"> Function to execute deconvolution upsample on OpenCL. - <td rowspan="1"> - <ul> - <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D - </ul> - <td>CLDeconvolutionLayerUpsample - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">DepthConvertLayer - <td rowspan="2" style="width:200px;"> Performs a down-scaling depth conversion. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEDepthConvertLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>F16, F32 - <tr><td>U8<td>U16, S16, S32 - <tr><td>U16<td>U8, U32 - <tr><td>S16<td>U8, S32 - <tr><td>BFLOAT16<td>F32 - <tr><td>F16<td>QASYMM8, F32 - <tr><td>F32<td>QASYMM8, F16, BFLOAT16 - </table> -<tr> - <td>CLDepthConvertLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>S8, U16, S16, U32, S32, F16, F32 - <tr><td>U16<td>U8, S8, S16, U32, S32, F16, F32 - <tr><td>S16<td>U8, S8, U16, U32, S32, F16, F32 - <tr><td>U32<td>U8, S8, U16, S16, S32, F16, F32 - <tr><td>S32<td>U8, S8, U16, S16, U32, F16, F32 - <tr><td>F16<td>U8, S8, U16, S16, U32, F32 - <tr><td>F32<td>U8, S8, U16, S16, U32, F16 - </table> -<tr> - <td rowspan="2">DepthToSpaceLayer - <td rowspan="2" style="width:200px;"> Depth to Space transformation. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_DEPTH_TO_SPACE - </ul> - <td>NEDepthToSpaceLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLDepthToSpaceLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">DepthwiseConvolutionLayer - <td rowspan="2" style="width:200px;"> Function to perform depthwise separable convolution. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_DEPTHWISE_CONV_2D - </ul> - <td>NEDepthwiseConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td>CLDepthwiseConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="2">DequantizationLayer - <td rowspan="2" style="width:200px;"> Function to dequantize the values in a tensor. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_DEQUANTIZE - </ul> - <td>NEDequantizationLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>F16, F32 - <tr><td>QASYMM8_SIGNED<td>F16, F32 - <tr><td>QSYMM8_PER_CHANNEL<td>F16, F32 - <tr><td>QSYMM8<td>F16, F32 - <tr><td>QSYMM16<td>F16, F32 - </table> -<tr> - <td>CLDequantizationLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>F16, F32 - <tr><td>QASYMM8_SIGNED<td>F16, F32 - <tr><td>QSYMM8_PER_CHANNEL<td>F16, F32 - <tr><td>QSYMM8<td>F16, F32 - <tr><td>QSYMM16<td>F16, F32 - </table> -<tr> - <td rowspan="1">DetectionPostProcessLayer - <td rowspan="1" style="width:200px;"> Function to generate the detection output based on center size encoded boxes, class prediction and anchors by doing non maximum suppression (NMS). - <td rowspan="1"> - <ul> - <li>ANEURALNETWORKS_DETECTION_POSTPROCESSING - </ul> - <td>NEDetectionPostProcessLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0 - src2<th>dst0 - dst3 - <tr><td>QASYMM8<td>F32 - <tr><td>QASYMM8_SIGNED<td>F32 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">DirectConvolutionLayer - <td rowspan="2" style="width:200px;"> Function to compute direct convolution. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_CONV_2D - </ul> - <td>NEDirectConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLDirectConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="1">DirectDeconvolutionLayer - <td rowspan="1" style="width:200px;"> Function to run the deconvolution layer. - <td rowspan="1"> - <ul> - <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D - </ul> - <td>CLDirectDeconvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="13">ElementWiseOperations - <td rowspan="13" style="width:200px;"> Function to perform in Cpu: - Div - Max - Min - Pow - SquaredDiff - Comparisons (Equal, greater, greater_equal, less, less_equal, not_equal) Function to perform in CL: - Add - Sub - Div - Max - Min - Pow - SquaredDiff - <td rowspan="13"> - <ul> - <li>ANEURALNETWORKS_MAXIMUM - <li>ANEURALNETWORKS_MINIMUM - <li>ANEURALNETWORKS_POW - <li>ANEURALNETWORKS_DIV - <li>ANEURALNETWORKS_ADD - <li>ANEURALNETWORKS_SUB - <li>ANEURALNETWORKS_EQUAL - <li>ANEURALNETWORKS_GREATER - <li>ANEURALNETWORKS_GREATER_EQUAL - <li>ANEURALNETWORKS_LESS - <li>ANEURALNETWORKS_LESS_EQUAL - <li>ANEURALNETWORKS_NOT_EQUAL - </ul> - <td>NEElementwiseMax - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>S32<td>S32<td>S32 - <tr><td>S16<td>S16<td>S16 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>NEElementwiseMin - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>S32<td>S32<td>S32 - <tr><td>S16<td>S16<td>S16 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>NEElementwiseSquaredDiff - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>S32<td>S32<td>S32 - <tr><td>S16<td>S16<td>S16 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>NEElementwiseDivision - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>NEElementwisePower - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>NEElementwiseComparison - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>U8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>U8 - <tr><td>S32<td>S32<td>U8 - <tr><td>U8<td>U8<td>U8 - <tr><td>S16<td>S16<td>U8 - <tr><td>F16<td>F16<td>U8 - <tr><td>F32<td>F32<td>U8 - </table> -<tr> - <td>CLArithmeticAddition - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16 - <tr><td>U8<td>U8<td>U8 - <tr><td>U8<td>U8<td>S16 - <tr><td>U8<td>S16<td>S16 - <tr><td>S16<td>U8<td>S16 - <tr><td>S16<td>S16<td>S16 - <tr><td>S32<td>S32<td>S32 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLArithmeticSubtraction - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16 - <tr><td>U8<td>U8<td>U8 - <tr><td>U8<td>U8<td>S16 - <tr><td>U8<td>S16<td>S16 - <tr><td>S16<td>U8<td>S16 - <tr><td>S16<td>S16<td>S16 - <tr><td>S32<td>S32<td>S32 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLArithmeticDivision - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLElementwiseMax - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16 - <tr><td>U8<td>U8<td>U8 - <tr><td>S16<td>S16<td>S16 - <tr><td>S32<td>S32<td>S32 - <tr><td>U32<td>U32<td>U32 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLElementwiseMin - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16 - <tr><td>U8<td>U8<td>U8 - <tr><td>S16<td>S16<td>S16 - <tr><td>S32<td>S32<td>S32 - <tr><td>U32<td>U32<td>U32 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLElementwiseSquaredDiff - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16 - <tr><td>U8<td>U8<td>U8 - <tr><td>S16<td>S16<td>S16 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLElementwisePower - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td rowspan="8">ElementwiseUnaryLayer - <td rowspan="8" style="width:200px;"> Function to perform: - Rsqrt - Exp - Neg - Log - Abs - Round - Sin - <td rowspan="8"> - <ul> - <li>ANEURALNETWORKS_ABS - <li>ANEURALNETWORKS_EXP - <li>ANEURALNETWORKS_LOG - <li>ANEURALNETWORKS_NEG - <li>ANEURALNETWORKS_RSQRT - <li>ANEURALNETWORKS_SIN - </ul> - <td>NEElementwiseUnaryLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - <tr><td>S32<td>S32 - </table> -<tr> - <td>CLRsqrtLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLExpLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLNegLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLSinLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLLogLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLAbsLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLRoundLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">FFT1D - <td rowspan="2" style="width:200px;"> Fast Fourier Transform 1D. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEFFT1D - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLFFT1D - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td rowspan="2">FFT2D - <td rowspan="2" style="width:200px;"> Fast Fourier Transform 2D. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEFFT2D - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLFFT2D - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td rowspan="2">FFTConvolutionLayer - <td rowspan="2" style="width:200px;"> Fast Fourier Transform Convolution. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_CONV_2D - </ul> - <td>NEFFTConvolutionLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLFFTConvolutionLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td rowspan="2">Fill - <td rowspan="2" style="width:200px;"> Set the values of a tensor with a given value. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_FILL - </ul> - <td>NEFill - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLFill - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">FillBorder - <td rowspan="2" style="width:200px;"> Function to . - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEFillBorder - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLFillBorder - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">FlattenLayer - <td rowspan="2" style="width:200px;"> Reshape a tensor to be 1D - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_RESHAPE - </ul> - <td>NEFlattenLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLFlattenLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">Floor - <td rowspan="2" style="width:200px;"> Round the value to the lowest number. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_FLOOR - </ul> - <td>NEFloor - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td>CLFloor - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td rowspan="2">FullyConnectedLayer - <td rowspan="2" style="width:200px;"> Function to perform a fully connected / dense layer. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_FULLY_CONNECTED - </ul> - <td>NEFullyConnectedLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td>CLFullyConnectedLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="2">FuseBatchNormalization - <td rowspan="2" style="width:200px;"> Function to fuse the batch normalization node to a preceding convolution node. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEFuseBatchNormalization - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td>CLFuseBatchNormalization - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td rowspan="2">Gather - <td rowspan="2" style="width:200px;"> Performs the Gather operation along the chosen axis. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_GATHER - </ul> - <td>NEGather - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLGather - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">GEMM - <td rowspan="2" style="width:200px;"> General Matrix Multiplication. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEGEMM - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16<td>BFLOAT16 - </table> -<tr> - <td>CLGEMM - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>F16<td>F16<td>F16<td>F16 - </table> -<tr> - <td rowspan="1">GEMMConv2D - <td rowspan="1" style="width:200px;"> General Matrix Multiplication. - <td rowspan="1"> - <ul> - <li>ANEURALNETWORKS_CONV_2D - </ul> - <td>NEGEMMConv2d - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16<td>BFLOAT16 - </table> -<tr> - <td rowspan="2">GEMMConvolutionLayer - <td rowspan="2" style="width:200px;"> General Matrix Multiplication. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_CONV_2D - </ul> - <td>NEGEMMConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16<td>BFLOAT16 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td>CLGEMMConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="1">GEMMDeconvolutionLayer - <td rowspan="1" style="width:200px;"> General Matrix Multiplication. - <td rowspan="1"> - <ul> - <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D - </ul> - <td>CLGEMMDeconvolutionLayer - <td> - <ul> - <li>NHWC - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="2">GEMMLowpMatrixMultiplyCore - <td rowspan="2" style="width:200px;"> General Matrix Multiplication. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEGEMMLowpMatrixMultiplyCore - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>S32 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>S32 - <tr><td>QASYMM8<td>QSYMM8<td>S32<td>S32 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>S32 - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>S32 - <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>S32 - </table> -<tr> - <td>CLGEMMLowpMatrixMultiplyCore - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QSYMM8<td>S32<td>QASYMM8 - <tr><td>QASYMM8<td>QASYMM8<td>S32<td>S32 - <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>S32 - <tr><td>QASYMM8<td>QSYMM8<td>S32<td>S32 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>QASYMM8_SIGNED - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>S32 - <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>S32 - <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>S32 - </table> -<tr> - <td rowspan="2">GEMMLowpOutputStage - <td rowspan="2" style="width:200px;"> General Matrix Multiplication. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEGEMMLowpOutputStage - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>S32<td>S32<td>QASYMM8 - <tr><td>S32<td>S32<td>QASYMM8_SIGNED - <tr><td>S32<td>S32<td>QSYMM16 - </table> -<tr> - <td>CLGEMMLowpOutputStage - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>S32<td>S32<td>QASYMM8 - <tr><td>S32<td>S32<td>QASYMM8_SIGNED - <tr><td>S32<td>S32<td>QSYMM16 - </table> -<tr> - <td rowspan="2">GenerateProposalsLayer - <td rowspan="2" style="width:200px;"> Function to generate proposals for a RPN (Region Proposal Network). - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_GENERATE_PROPOSALS - </ul> - <td>NEGenerateProposalsLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QSYMM8<td>QSYMM16<td>QASYMM8 - </table> -<tr> - <td>CLGenerateProposalsLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QSYMM8<td>QSYMM16<td>QASYMM8 - </table> -<tr> - <td rowspan="2">InstanceNormalizationLayer - <td rowspan="2" style="width:200px;"> Function to perform a Instance normalization on a given axis. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_INSTANCE_NORMALIZATION - </ul> - <td>NEInstanceNormalizationLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLInstanceNormalizationLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">L2NormalizeLayer - <td rowspan="2" style="width:200px;"> Function to perform a L2 normalization on a given axis. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_L2_NORMALIZATION - </ul> - <td>NEL2NormalizeLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLL2NormalizeLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="3">Logical - <td rowspan="3" style="width:200px;"> Function to perform: - Logical AND - Logical OR - Logical NOT - <td rowspan="3"> - <ul> - <li>n/a - </ul> - <td>NELogicalAnd - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>U8<td>U8<td>U8 - </table> -<tr> - <td>NELogicalOr - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>U8<td>U8<td>U8 - </table> -<tr> - <td>NELogicalNot - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td rowspan="1">LogicalAnd - <td rowspan="1" style="width:200px;"> Function to perform Logical AND. - <td rowspan="1"> - <ul> - <li>n/a - </ul> - <td>CLLogicalAnd - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>U8<td>U8<td>U8 - </table> -<tr> - <td rowspan="1">LogicalOr - <td rowspan="1" style="width:200px;"> Function to perform Logical OR. - <td rowspan="1"> - <ul> - <li>n/a - </ul> - <td>CLLogicalOr - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>U8<td>U8<td>U8 - </table> -<tr> - <td rowspan="1">LogicalNot - <td rowspan="1" style="width:200px;"> Function to perform Logical NOT. - <td rowspan="1"> - <ul> - <li>n/a - </ul> - <td>CLLogicalNot - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>U8<td>U8 - </table> -<tr> - <td rowspan="2">LSTMLayer - <td rowspan="2" style="width:200px;"> Function to perform a single time step in a Long Short-Term Memory (LSTM) layer. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_LSTM - </ul> - <td>NELSTMLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0 - src13<th>dst0 - dst3 - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLLSTMLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0 - src13<th>dst0 - dst3 - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">LSTMLayerQuantized - <td rowspan="2" style="width:200px;"> Function to perform quantized LSTM (Long Short-Term Memory) - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_QUANTIZED_LSTM - <li>ANEURALNETWORKS_QUANTIZED_16BIT_LSTM - </ul> - <td>NELSTMLayerQuantized - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0 - src8<th>src9 - src12<th>src13<th>src14<th>dst0<th>dst1 - <tr><td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8<td>QSYMM16<td>QASYMM8 - </table> -<tr> - <td>CLLSTMLayerQuantized - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0 - src8<th>src9 - src12<th>src13<th>src14<th>dst0<th>dst1 - <tr><td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8<td>QSYMM16<td>QASYMM8 - </table> -<tr> - <td rowspan="2">MaxUnpoolingLayer - <td rowspan="2" style="width:200px;"> Function to perform MaxUnpooling. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEMaxUnpoolingLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLMaxUnpoolingLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">MeanStdDevNormalizationLayer - <td rowspan="2" style="width:200px;"> Function to execute mean and standard deviation normalization. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEMeanStdDevNormalizationLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td>CLMeanStdDevNormalizationLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td rowspan="2">NormalizationLayer - <td rowspan="2" style="width:200px;"> Function to compute normalization layer. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_LOCAL_RESPONSE_NORMALIZATION - </ul> - <td>NENormalizationLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td>CLNormalizationLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F32<td>F32 - <tr><td>F16<td>F16 - </table> -<tr> - <td rowspan="2">PadLayer - <td rowspan="2" style="width:200px;"> Function to pad a tensor. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_PAD - <li>ANEURALNETWORKS_PAD_V2 - </ul> - <td>NEPadLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLPadLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">Permute - <td rowspan="2" style="width:200px;"> Function to transpose an ND tensor. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_TRANSPOSE - </ul> - <td>NEPermute - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLPermute - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">PixelWiseMultiplication - <td rowspan="2" style="width:200px;"> Function to performe a multiplication. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_MUL - </ul> - <td>NEPixelWiseMultiplication - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16 - <tr><td>QSYMM16<td>QSYMM16<td>S32 - <tr><td>U8<td>U8<td>U8 - <tr><td>U8<td>U8<td>S16 - <tr><td>U8<td>S16<td>S16 - <tr><td>S16<td>U8<td>S16 - <tr><td>S16<td>S16<td>S16 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>S32<td>F32 - </table> -<tr> - <td>CLPixelWiseMultiplication - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16 - <tr><td>QSYMM16<td>QSYMM16<td>S32 - <tr><td>U8<td>U8<td>U8 - <tr><td>U8<td>U8<td>S16 - <tr><td>U8<td>S16<td>S16 - <tr><td>S16<td>U8<td>S16 - <tr><td>S16<td>S16<td>S16 - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>S32<td>F32 - </table> -<tr> - <td rowspan="2">PoolingLayer - <td rowspan="2" style="width:200px;"> Function to performe pooling with the specified pooling operation. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_AVERAGE_POOL_2D - <li>ANEURALNETWORKS_L2_POOL_2D - <li>ANEURALNETWORKS_MAX_POOL_2D - </ul> - <td>NEPoolingLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLPoolingLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">PReluLayer - <td rowspan="2" style="width:200px;"> Function to compute the activation layer with the PRELU activation function. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_PRELU - </ul> - <td>NEPReluLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLPReluLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">PriorBoxLayer - <td rowspan="2" style="width:200px;"> Function to compute prior boxes and clip. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEPriorBoxLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLPriorBoxLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F32<td>F32<td>F32 - </table> -<tr> - <td rowspan="2">QLSTMLayer - <td rowspan="2" style="width:200px;"> Function to perform quantized LSTM (Long Short-Term Memory). - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_QUANTIZED_LSTM - <li>ANEURALNETWORKS_QUANTIZED_16BIT_LSTM - </ul> - <td>NEQLSTMLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1 - src6<th>src7 -src9<th>src10<th>src11<th>dst0<th>dst1 - dst2 - <tr><td>QASYMM8_SIGNED<td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8_SIGNED<td>QSYMM16<td>QASYMM8_SIGNED - </table> -<tr> - <td>CLQLSTMLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1 - src6<th>src7 -src9<th>src10<th>src11<th>dst0<th>dst1 - dst2 - <tr><td>QASYMM8_SIGNED<td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8_SIGNED<td>QSYMM16<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="2">QuantizationLayer - <td rowspan="2" style="width:200px;"> Function to perform quantization layer - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_QUANTIZE - </ul> - <td>NEQuantizationLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8, QASYMM8_SIGNED, QASYMM16 - <tr><td>QASYMM8_SIGNED<td>QASYMM8, QASYMM8_SIGNED, QASYMM16 - <tr><td>F16<td>QASYMM8, QASYMM8_SIGNED, QASYMM16 - <tr><td>F32<td>QASYMM8, QASYMM8_SIGNED, QASYMM16 - </table> -<tr> - <td>CLQuantizationLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8, QASYMM8_SIGNED, QASYMM16 - <tr><td>QASYMM8_SIGNED<td>QASYMM8, QASYMM8_SIGNED, QASYMM16 - <tr><td>F16<td>QASYMM8, QASYMM8_SIGNED, QASYMM16 - <tr><td>F32<td>QASYMM8, QASYMM8_SIGNED, QASYMM16 - </table> -<tr> - <td rowspan="2">Range - <td rowspan="2" style="width:200px;"> Function to generates a sequence of numbers starting from START and extends by increments of 'STEP' up to but not including 'END'. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NERange - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>dst - <tr><td>U8 - <tr><td>S8 - <tr><td>U16 - <tr><td>S16 - <tr><td>U32 - <tr><td>S32 - <tr><td>F16 - <tr><td>F32 - </table> -<tr> - <td>CLRange - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>dst - <tr><td>U8 - <tr><td>S8 - <tr><td>QASYMM8 - <tr><td>U16 - <tr><td>S16 - <tr><td>U32 - <tr><td>S32 - <tr><td>F16 - <tr><td>F32 - </table> -<tr> - <td rowspan="2">ReduceMean - <td rowspan="2" style="width:200px;"> Function to performe reduce mean operation. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_MEAN - </ul> - <td>NEReduceMean - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLReduceMean - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">ReductionOperation - <td rowspan="2" style="width:200px;"> Function to performe reduce with the following operations - ARG_IDX_MAX: Index of the max value - ARG_IDX_MIN: Index of the min value - MEAN_SUM: Mean of sum - PROD: Product - SUM_SQUARE: Sum of squares - SUM: Sum - MIN: Min - MAX: Max - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_REDUCE_ALL - <li>ANEURALNETWORKS_REDUCE_ANY - <li>ANEURALNETWORKS_REDUCE_MAX - <li>ANEURALNETWORKS_REDUCE_MIN - <li>ANEURALNETWORKS_REDUCE_PROD - <li>ANEURALNETWORKS_REDUCE_SUM - </ul> - <td>NEReductionOperation - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - <tr><td>S32<td>S32 - </table> -<tr> - <td>CLReductionOperation - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - <tr><td>S32<td>S32 - </table> -<tr> - <td rowspan="2">ReorgLayer - <td rowspan="2" style="width:200px;"> Performs a reorganization layer of input tensor to the output tensor. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEReorgLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLReorgLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">ReshapeLayer - <td rowspan="2" style="width:200px;"> Function to reshape a tensor. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_RESHAPE - <li>ANEURALNETWORKS_SQUEEZE - </ul> - <td>NEReshapeLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLReshapeLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">Reverse - <td rowspan="2" style="width:200px;"> Function to reverse tensor according to axis. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEReverse - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>All<td>U32<td>All - </table> -<tr> - <td>CLReverse - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>All<td>U32<td>All - </table> -<tr> - <td rowspan="2">RNNLayer - <td rowspan="2" style="width:200px;"> Function to perform recurrent neural network layer. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_RNN - </ul> - <td>NERNNLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>src3<th>dst0<th>dst1 - <tr><td>F16<td>F16<td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLRNNLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>src3<th>dst0<th>dst1 - <tr><td>F16<td>F16<td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32<td>F32<td>F32 - </table> -<tr> - <td rowspan="2">ROIAlignLayer - <td rowspan="2" style="width:200px;"> Function to perform ROI alignment. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_ROI_ALIGN - </ul> - <td>NEROIAlignLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM16<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM16<td>QASYMM8_SIGNED - </table> -<tr> - <td>CLROIAlignLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32 - <tr><td>QASYMM8<td>QASYMM16<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM16<td>QASYMM8_SIGNED - </table> -<tr> - <td rowspan="2">ROIPoolingLayer - <td rowspan="2" style="width:200px;"> Function to perform ROI pooling. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_ROI_POOLING - </ul> - <td>NEROIPoolingLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F32<td>U16<td>F32 - <tr><td>QASYMM8<td>U16<td>QASYMM8 - </table> -<tr> - <td>CLROIPoolingLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>dst - <tr><td>F16<td>U16<td>F16 - <tr><td>F32<td>U16<td>F32 - <tr><td>QASYMM8<td>U16<td>QASYMM8 - </table> -<tr> - <td rowspan="2">Scale - <td rowspan="2" style="width:200px;"> Function to perform resize a tensor using to interpolate: - Bilinear - Nearest neighbor - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_RESIZE_BILINEAR - <li>ANEURALNETWORKS_RESIZE_NEAREST_NEIGHBOR - </ul> - <td>NEScale - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - <tr><td>U8<td>U8 - <tr><td>S16<td>S16 - </table> -<tr> - <td>CLScale - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - <tr><td>U8<td>U8 - <tr><td>S16<td>S16 - </table> -<tr> - <td rowspan="2">Select - <td rowspan="2" style="width:200px;"> Function to select values from 2 tensors depending on an input tensor of booleans. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_SELECT - </ul> - <td>NESelect - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>U8<td>All<td>All<td>All - </table> -<tr> - <td>CLSelect - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>U8<td>All<td>All<td>All - </table> -<tr> - <td rowspan="2">Slice - <td rowspan="2" style="width:200px;"> Function to perform tensor slicing. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_SLICE - </ul> - <td>NESlice - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLSlice - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">SoftmaxLayer - <td rowspan="2" style="width:200px;"> Function to compute a SoftmaxLayer and a Log SoftmaxLayer. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_LOG_SOFTMAX - <li>ANEURALNETWORKS_SOFTMAX - </ul> - <td>NESoftmaxLayerGeneric - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td>CLSoftmaxLayerGeneric - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>QASYMM8<td>QASYMM8 - <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -<tr> - <td rowspan="2">SpaceToBatchLayer - <td rowspan="2" style="width:200px;"> Function to divide a tensor spatially. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_SPACE_TO_BATCH_ND - </ul> - <td>NESpaceToBatchLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>All<td>S32<td>S32<td>All - </table> -<tr> - <td>CLSpaceToBatchLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>All<td>S32<td>S32<td>All - </table> -<tr> - <td rowspan="2">SpaceToDepthLayer - <td rowspan="2" style="width:200px;"> Function to rearrange blocks of spatial data into depth. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_SPACE_TO_DEPTH - </ul> - <td>NESpaceToDepthLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLSpaceToDepthLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">Split - <td rowspan="2" style="width:200px;"> Function to split a tensor along a given axis. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_SPLIT - </ul> - <td>NESplit - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLSplit - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">StackLayer - <td rowspan="2" style="width:200px;"> Function to stack tensors along an axis. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEStackLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLStackLayer - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">StridedSlice - <td rowspan="2" style="width:200px;"> Function to extract a strided slice of a tensor. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_STRIDED_SLICE - </ul> - <td>NEStridedSlice - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLStridedSlice - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">Tile - <td rowspan="2" style="width:200px;"> Function to construct a tensor by tiling a given tensor. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_TILE - </ul> - <td>NETile - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLTile - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">Transpose - <td rowspan="2" style="width:200px;"> Function to transpose a 2D tensor. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_TRANSPOSE - </ul> - <td>NETranspose - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLTranspose - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">Unstack - <td rowspan="2" style="width:200px;"> Function to unpack a rank-R tensor into rank-(R-1) tensors. - <td rowspan="2"> - <ul> - <li>n/a - </ul> - <td>NEUnstack - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td>CLUnstack - <td> - <ul> - <li>All - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>All<td>All - </table> -<tr> - <td rowspan="2">WinogradConvolutionLayer - <td rowspan="2" style="width:200px;"> Function to do Winograd Convolution. - <td rowspan="2"> - <ul> - <li>ANEURALNETWORKS_CONV_2D - </ul> - <td>NEWinogradConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - </table> -<tr> - <td>CLWinogradConvolutionLayer - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src0<th>src1<th>src2<th>dst - <tr><td>F16<td>F16<td>F16<td>F16 - <tr><td>F32<td>F32<td>F32<td>F32 - </table> -<tr> - <td rowspan="1">WinogradInputTransform - <td rowspan="1" style="width:200px;"> Function to. - <td rowspan="1"> - <ul> - <li>n/a - </ul> - <td>CLWinogradInputTransform - <td> - <ul> - <li>NHWC - <li>NCHW - </ul> - <td> - <table> - <tr><th>src<th>dst - <tr><td>F16<td>F16 - <tr><td>F32<td>F32 - </table> -</table> - -*/ -} // namespace
\ No newline at end of file |