aboutsummaryrefslogtreecommitdiff
path: root/docs/09_operators_list.dox
diff options
context:
space:
mode:
authorSang-Hoon Park <sang-hoon.park@arm.com>2021-05-05 10:34:47 +0100
committerSheri Zhang <sheri.zhang@arm.com>2021-05-05 11:43:47 +0000
commitc9309f22a026dfce92365e2f0802c40e8e1c449e (patch)
treec3d1b8498bb319d8716dd78979b24fd24ae8c411 /docs/09_operators_list.dox
parentd813bab10bb4fe954fa0e962e1402ed1377617da (diff)
downloadComputeLibrary-c9309f22a026dfce92365e2f0802c40e8e1c449e.tar.gz
Restructure Documentation (Part 2)
The followings are done: - Move operator list documnetation - Introduction page is moved to the top - The sections for experimental API and programming model are merged into library architecture page. Resolves: COMPMID-4198 Change-Id: Iad824d6c8ba8d31e0bf76afd3fb67abbe32a1667 Signed-off-by: Sang-Hoon Park <sang-hoon.park@arm.com> Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/5570 Tested-by: Arm Jenkins <bsgcomp@arm.com> Reviewed-by: Michele Di Giorgio <michele.digiorgio@arm.com> Reviewed-by: Sheri Zhang <sheri.zhang@arm.com> Comments-Addressed: Arm Jenkins <bsgcomp@arm.com>
Diffstat (limited to 'docs/09_operators_list.dox')
-rw-r--r--docs/09_operators_list.dox3123
1 files changed, 0 insertions, 3123 deletions
diff --git a/docs/09_operators_list.dox b/docs/09_operators_list.dox
deleted file mode 100644
index fc41265738..0000000000
--- a/docs/09_operators_list.dox
+++ /dev/null
@@ -1,3123 +0,0 @@
-///
-/// Copyright (c) 2021 Arm Limited.
-///
-/// SPDX-License-Identifier: MIT
-///
-/// Permission is hereby granted, free of charge, to any person obtaining a copy
-/// of this software and associated documentation files (the "Software"), to
-/// deal in the Software without restriction, including without limitation the
-/// rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
-/// sell copies of the Software, and to permit persons to whom the Software is
-/// furnished to do so, subject to the following conditions:
-///
-/// The above copyright notice and this permission notice shall be included in all
-/// copies or substantial portions of the Software.
-///
-/// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
-/// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
-/// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
-/// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
-/// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
-/// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
-/// SOFTWARE.
-///
-namespace arm_compute
-{
-/**
-@page operators_list Supported Operators
-
-@tableofcontents
-
-@section S9_1_operators_list Supported Operators
-
-Compute Library supports operators that are listed in below table.
-
-Compute Library supports a wide list of data-types, information can been directly found in the documentation of each kernel/function.
-The main data-types that the Machine Learning functions support are the following:
- <ul>
- <li>BFLOAT16: 16-bit non-standard brain floating point
- <li>QASYMM8: 8-bit unsigned asymmetric quantized
- <li>QASYMM8_SIGNED: 8-bit signed asymmetric quantized
- <li>QSYMM8_PER_CHANNEL: 8-bit signed symmetric quantized (Used for the weights)
- <li>QSYMM8: 8-bit unsigned symmetric quantized
- <li>QSYMM16: 16-bit unsigned symmetric quantized
- <li>F32: 32-bit single precision floating point
- <li>F16: 16-bit half precision floating point
- <li>S32: 32-bit signed integer
- <li>U8: 8-bit unsigned char
- <li>All: include all above data types
- </ul>
-
-Compute Library supports the following data layouts (fast changing dimension from right to left):
- <ul>
- <li>NHWC: The native layout of Compute Library that delivers the best performance where channels are in the fastest changing dimension
- <li>NCHW: Legacy layout where width is in the fastest changing dimension
- <li>All: include all above data layouts
- </ul>
-where N = batches, C = channels, H = height, W = width
-
-<table>
-<caption id="multi_row"></caption>
-<tr>
- <th>Function
- <th>Description
- <th>Equivalent Android NNAPI Op
- <th>Backends
- <th>Data Layouts
- <th>Data Types
-<tr>
- <td rowspan="2">ActivationLayer
- <td rowspan="2" style="width:200px;"> Function to simulate an activation layer with the specified activation function.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_ELU
- <li>ANEURALNETWORKS_HARD_SWISH
- <li>ANEURALNETWORKS_LOGISTIC
- <li>ANEURALNETWORKS_RELU
- <li>ANEURALNETWORKS_RELU1
- <li>ANEURALNETWORKS_RELU6
- <li>ANEURALNETWORKS_TANH
- </ul>
- <td>NEActivationLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLActivationLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">ArgMinMaxLayer
- <td rowspan="2" style="width:200px;"> Function to calculate the index of the minimum or maximum values in a tensor based on an axis.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_ARGMAX
- <li>ANEURALNETWORKS_ARGMIN
- </ul>
- <td>NEArgMinMaxLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>U32, S32
- <tr><td>QASYMM8_SIGNED<td>U32, S32
- <tr><td>S32<td>U32, S32
- <tr><td>F16<td>U32, S32
- <tr><td>F32<td>U32, S32
- </table>
-<tr>
- <td>CLArgMinMaxLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>U32, S32
- <tr><td>QASYMM8_SIGNED<td>U32, S32
- <tr><td>S32<td>U32, S32
- <tr><td>F16<td>U32, S32
- <tr><td>F32<td>U32, S32
- </table>
-<tr>
- <td rowspan="1">ArithmeticAddition
- <td rowspan="1" style="width:200px;"> Function to add 2 tensors.
- <td rowspan="1">
- <ul>
- <li>ANEURALNETWORKS_ADD
- </ul>
- <td>NEArithmeticAddition
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
- <tr><td>QSYMM16<td>QSYMM16<td>S32
- <tr><td>U8<td>U8<td>U8
- <tr><td>U8<td>U8<td>S16
- <tr><td>U8<td>S16<td>S16
- <tr><td>S16<td>U8<td>S16
- <tr><td>S16<td>S16<td>S16
- <tr><td>S32<td>S32<td>S32
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td rowspan="1">ArithmeticSubtraction
- <td rowspan="1" style="width:200px;"> Function to substract 2 tensors.
- <td rowspan="1">
- <ul>
- <li>ANEURALNETWORKS_SUB
- </ul>
- <td>NEArithmeticSubtraction
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
- <tr><td>QSYMM16<td>QSYMM16<td>S32
- <tr><td>U8<td>U8<td>U8
- <tr><td>U8<td>U8<td>S16
- <tr><td>U8<td>S16<td>S16
- <tr><td>S16<td>U8<td>S16
- <tr><td>S16<td>S16<td>S16
- <tr><td>S32<td>S32<td>S32
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">BatchNormalizationLayer
- <td rowspan="2" style="width:200px;"> Function to perform batch normalization.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEBatchNormalizationLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td>CLBatchNormalizationLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td rowspan="2">BatchToSpaceLayer
- <td rowspan="2" style="width:200px;"> Batch to space transformation.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_BATCH_TO_SPACE_ND
- </ul>
- <td>NEBatchToSpaceLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>All<td>s32<td>All
- </table>
-<tr>
- <td>CLBatchToSpaceLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>All<td>s32<td>All
- </table>
-<tr>
- <td rowspan="2">BitwiseAnd
- <td rowspan="2" style="width:200px;"> Function to performe bitwise AND between 2 tensors.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_LOGICAL_AND
- </ul>
- <td>NEBitwiseAnd
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td>CLBitwiseAnd
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td rowspan="2">BitwiseNot
- <td rowspan="2" style="width:200px;"> Function to performe bitwise NOT.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_LOGICAL_NOT
- </ul>
- <td>NEBitwiseNot
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td>CLBitwiseNot
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td rowspan="2">BitwiseOr
- <td rowspan="2" style="width:200px;"> Function to performe bitwise OR between 2 tensors.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_LOGICAL_OR
- </ul>
- <td>NEBitwiseOr
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td>CLBitwiseOr
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td rowspan="2">BitwiseXor
- <td rowspan="2" style="width:200px;"> Function to performe bitwise XOR between 2 tensors.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEBitwiseXor
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td>CLBitwiseXor
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td rowspan="2">BoundingBoxTransform
- <td rowspan="2" style="width:200px;"> Transform proposal bounding boxes to target bounding box using bounding box deltas.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEBoundingBoxTransform
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM16<td>QASYMM8<td>QASYMM16
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLBoundingBoxTransform
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM16<td>QASYMM8<td>QASYMM16
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">Cast
- <td rowspan="2" style="width:200px;"> Function to cast a tensor.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_CAST
- </ul>
- <td>NECast
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8_SIGNED<td>S16, S32, F32, F16
- <tr><td>QASYMM8<td>U16, S16, S32, F32, F16
- <tr><td>U8<td>U16, S16, S32, F32, F16
- <tr><td>U16<td>U8, U32
- <tr><td>S16<td>QASYMM8_SIGNED, U8, S32
- <tr><td>F16<td>QASYMM8_SIGNED, QASYMM8, F32, S32, U8
- <tr><td>S32<td>QASYMM8_SIGNED, QASYMM8, F16, F32, U8
- <tr><td>F32<td>QASYMM8_SIGNED, QASYMM8, BFLOAT16, F16, S32, U8
- </table>
-<tr>
- <td>CLCast
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>S8, U16, S16, U32, S32, F16, F32
- <tr><td>U16<td>U8, S8, S16, U32, S32, F16, F32
- <tr><td>S16<td>U8, S8, U16, U32, S32, F16, F32
- <tr><td>U32<td>U8, S8, U16, S16, S32, F16, F32
- <tr><td>S32<td>U8, S8, U16, S16, U32, F16, F32
- <tr><td>F16<td>U8, S8, U16, S16, U32, F32
- <tr><td>F32<td>U8, S8, U16, S16, U32, F16
- </table>
-<tr>
- <td rowspan="2">ChannelShuffleLayer
- <td rowspan="2" style="width:200px;"> Function to shuffle the channels of the input tensor.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_CHANNEL_SHUFFLE
- </ul>
- <td>NEChannelShuffleLayer
- <td>
- <ul>
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLChannelShuffleLayer
- <td>
- <ul>
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="1">Comparison
- <td rowspan="1" style="width:200px;"> Function to compare 2 tensors.
- <td rowspan="1">
- <ul>
- <li>ANEURALNETWORKS_EQUAL
- <li>ANEURALNETWORKS_GREATER
- <li>ANEURALNETWORKS_GREATER_EQUAL
- <li>ANEURALNETWORKS_LESS
- <li>ANEURALNETWORKS_LESS_EQUAL
- <li>ANEURALNETWORKS_NOT_EQUAL
- </ul>
- <td>CLComparison
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>All<td>All<td>U8
- </table>
-<tr>
- <td rowspan="2">ConcatenateLayer
- <td rowspan="2" style="width:200px;"> Function to concatenate tensors along a given axis.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_CONCATENATION
- </ul>
- <td>NEConcatenateLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLConcatenateLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">ConvertFullyConnectedWeights
- <td rowspan="2" style="width:200px;"> Function to tranpose the wieghts for the fully connected layer.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEConvertFullyConnectedWeights
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLConvertFullyConnectedWeights
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">ConvolutionLayer
- <td rowspan="2" style="width:200px;"> Function to compute a convolution layer.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_CONV_2D
- </ul>
- <td>NEConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td>CLConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="2">Copy
- <td rowspan="2" style="width:200px;"> Function to copy a tensor.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NECopy
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLCopy
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="1">Crop
- <td rowspan="1" style="width:200px;"> Performs a copy of input tensor to the output tensor.
- <td rowspan="1">
- <ul>
- <li>n/a
- </ul>
- <td>CLCrop
- <td>
- <ul>
- <li>NHWC
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>F32
- </table>
-<tr>
- <td rowspan="2">CropResize
- <td rowspan="2" style="width:200px;"> Function to perform cropping and resizing.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NECropResize
- <td>
- <ul>
- <li>NHWC
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>All<td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLCropResize
- <td>
- <ul>
- <li>NHWC
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>All<td>F32<td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">DeconvolutionLayer
- <td rowspan="2" style="width:200px;"> Function to compute a deconvolution or tranpose convolution.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D
- </ul>
- <td>NEDeconvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td>CLDeconvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="1">DeconvolutionLayerUpsample
- <td rowspan="1" style="width:200px;"> Function to execute deconvolution upsample on OpenCL.
- <td rowspan="1">
- <ul>
- <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D
- </ul>
- <td>CLDeconvolutionLayerUpsample
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">DepthConvertLayer
- <td rowspan="2" style="width:200px;"> Performs a down-scaling depth conversion.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEDepthConvertLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>F16, F32
- <tr><td>U8<td>U16, S16, S32
- <tr><td>U16<td>U8, U32
- <tr><td>S16<td>U8, S32
- <tr><td>BFLOAT16<td>F32
- <tr><td>F16<td>QASYMM8, F32
- <tr><td>F32<td>QASYMM8, F16, BFLOAT16
- </table>
-<tr>
- <td>CLDepthConvertLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>S8, U16, S16, U32, S32, F16, F32
- <tr><td>U16<td>U8, S8, S16, U32, S32, F16, F32
- <tr><td>S16<td>U8, S8, U16, U32, S32, F16, F32
- <tr><td>U32<td>U8, S8, U16, S16, S32, F16, F32
- <tr><td>S32<td>U8, S8, U16, S16, U32, F16, F32
- <tr><td>F16<td>U8, S8, U16, S16, U32, F32
- <tr><td>F32<td>U8, S8, U16, S16, U32, F16
- </table>
-<tr>
- <td rowspan="2">DepthToSpaceLayer
- <td rowspan="2" style="width:200px;"> Depth to Space transformation.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_DEPTH_TO_SPACE
- </ul>
- <td>NEDepthToSpaceLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLDepthToSpaceLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">DepthwiseConvolutionLayer
- <td rowspan="2" style="width:200px;"> Function to perform depthwise separable convolution.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_DEPTHWISE_CONV_2D
- </ul>
- <td>NEDepthwiseConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td>CLDepthwiseConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="2">DequantizationLayer
- <td rowspan="2" style="width:200px;"> Function to dequantize the values in a tensor.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_DEQUANTIZE
- </ul>
- <td>NEDequantizationLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>F16, F32
- <tr><td>QASYMM8_SIGNED<td>F16, F32
- <tr><td>QSYMM8_PER_CHANNEL<td>F16, F32
- <tr><td>QSYMM8<td>F16, F32
- <tr><td>QSYMM16<td>F16, F32
- </table>
-<tr>
- <td>CLDequantizationLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>F16, F32
- <tr><td>QASYMM8_SIGNED<td>F16, F32
- <tr><td>QSYMM8_PER_CHANNEL<td>F16, F32
- <tr><td>QSYMM8<td>F16, F32
- <tr><td>QSYMM16<td>F16, F32
- </table>
-<tr>
- <td rowspan="1">DetectionPostProcessLayer
- <td rowspan="1" style="width:200px;"> Function to generate the detection output based on center size encoded boxes, class prediction and anchors by doing non maximum suppression (NMS).
- <td rowspan="1">
- <ul>
- <li>ANEURALNETWORKS_DETECTION_POSTPROCESSING
- </ul>
- <td>NEDetectionPostProcessLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0 - src2<th>dst0 - dst3
- <tr><td>QASYMM8<td>F32
- <tr><td>QASYMM8_SIGNED<td>F32
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">DirectConvolutionLayer
- <td rowspan="2" style="width:200px;"> Function to compute direct convolution.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_CONV_2D
- </ul>
- <td>NEDirectConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLDirectConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="1">DirectDeconvolutionLayer
- <td rowspan="1" style="width:200px;"> Function to run the deconvolution layer.
- <td rowspan="1">
- <ul>
- <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D
- </ul>
- <td>CLDirectDeconvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="13">ElementWiseOperations
- <td rowspan="13" style="width:200px;"> Function to perform in Cpu: - Div - Max - Min - Pow - SquaredDiff - Comparisons (Equal, greater, greater_equal, less, less_equal, not_equal) Function to perform in CL: - Add - Sub - Div - Max - Min - Pow - SquaredDiff
- <td rowspan="13">
- <ul>
- <li>ANEURALNETWORKS_MAXIMUM
- <li>ANEURALNETWORKS_MINIMUM
- <li>ANEURALNETWORKS_POW
- <li>ANEURALNETWORKS_DIV
- <li>ANEURALNETWORKS_ADD
- <li>ANEURALNETWORKS_SUB
- <li>ANEURALNETWORKS_EQUAL
- <li>ANEURALNETWORKS_GREATER
- <li>ANEURALNETWORKS_GREATER_EQUAL
- <li>ANEURALNETWORKS_LESS
- <li>ANEURALNETWORKS_LESS_EQUAL
- <li>ANEURALNETWORKS_NOT_EQUAL
- </ul>
- <td>NEElementwiseMax
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>S32<td>S32<td>S32
- <tr><td>S16<td>S16<td>S16
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>NEElementwiseMin
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>S32<td>S32<td>S32
- <tr><td>S16<td>S16<td>S16
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>NEElementwiseSquaredDiff
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>S32<td>S32<td>S32
- <tr><td>S16<td>S16<td>S16
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>NEElementwiseDivision
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>NEElementwisePower
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>NEElementwiseComparison
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>U8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>U8
- <tr><td>S32<td>S32<td>U8
- <tr><td>U8<td>U8<td>U8
- <tr><td>S16<td>S16<td>U8
- <tr><td>F16<td>F16<td>U8
- <tr><td>F32<td>F32<td>U8
- </table>
-<tr>
- <td>CLArithmeticAddition
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
- <tr><td>U8<td>U8<td>U8
- <tr><td>U8<td>U8<td>S16
- <tr><td>U8<td>S16<td>S16
- <tr><td>S16<td>U8<td>S16
- <tr><td>S16<td>S16<td>S16
- <tr><td>S32<td>S32<td>S32
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLArithmeticSubtraction
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
- <tr><td>U8<td>U8<td>U8
- <tr><td>U8<td>U8<td>S16
- <tr><td>U8<td>S16<td>S16
- <tr><td>S16<td>U8<td>S16
- <tr><td>S16<td>S16<td>S16
- <tr><td>S32<td>S32<td>S32
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLArithmeticDivision
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLElementwiseMax
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
- <tr><td>U8<td>U8<td>U8
- <tr><td>S16<td>S16<td>S16
- <tr><td>S32<td>S32<td>S32
- <tr><td>U32<td>U32<td>U32
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLElementwiseMin
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
- <tr><td>U8<td>U8<td>U8
- <tr><td>S16<td>S16<td>S16
- <tr><td>S32<td>S32<td>S32
- <tr><td>U32<td>U32<td>U32
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLElementwiseSquaredDiff
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
- <tr><td>U8<td>U8<td>U8
- <tr><td>S16<td>S16<td>S16
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLElementwisePower
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td rowspan="8">ElementwiseUnaryLayer
- <td rowspan="8" style="width:200px;"> Function to perform: - Rsqrt - Exp - Neg - Log - Abs - Round - Sin
- <td rowspan="8">
- <ul>
- <li>ANEURALNETWORKS_ABS
- <li>ANEURALNETWORKS_EXP
- <li>ANEURALNETWORKS_LOG
- <li>ANEURALNETWORKS_NEG
- <li>ANEURALNETWORKS_RSQRT
- <li>ANEURALNETWORKS_SIN
- </ul>
- <td>NEElementwiseUnaryLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- <tr><td>S32<td>S32
- </table>
-<tr>
- <td>CLRsqrtLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLExpLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLNegLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLSinLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLLogLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLAbsLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLRoundLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">FFT1D
- <td rowspan="2" style="width:200px;"> Fast Fourier Transform 1D.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEFFT1D
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLFFT1D
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td rowspan="2">FFT2D
- <td rowspan="2" style="width:200px;"> Fast Fourier Transform 2D.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEFFT2D
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLFFT2D
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td rowspan="2">FFTConvolutionLayer
- <td rowspan="2" style="width:200px;"> Fast Fourier Transform Convolution.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_CONV_2D
- </ul>
- <td>NEFFTConvolutionLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLFFTConvolutionLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td rowspan="2">Fill
- <td rowspan="2" style="width:200px;"> Set the values of a tensor with a given value.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_FILL
- </ul>
- <td>NEFill
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLFill
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">FillBorder
- <td rowspan="2" style="width:200px;"> Function to .
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEFillBorder
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLFillBorder
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">FlattenLayer
- <td rowspan="2" style="width:200px;"> Reshape a tensor to be 1D
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_RESHAPE
- </ul>
- <td>NEFlattenLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLFlattenLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">Floor
- <td rowspan="2" style="width:200px;"> Round the value to the lowest number.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_FLOOR
- </ul>
- <td>NEFloor
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td>CLFloor
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td rowspan="2">FullyConnectedLayer
- <td rowspan="2" style="width:200px;"> Function to perform a fully connected / dense layer.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_FULLY_CONNECTED
- </ul>
- <td>NEFullyConnectedLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td>CLFullyConnectedLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="2">FuseBatchNormalization
- <td rowspan="2" style="width:200px;"> Function to fuse the batch normalization node to a preceding convolution node.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEFuseBatchNormalization
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td>CLFuseBatchNormalization
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td rowspan="2">Gather
- <td rowspan="2" style="width:200px;"> Performs the Gather operation along the chosen axis.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_GATHER
- </ul>
- <td>NEGather
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLGather
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">GEMM
- <td rowspan="2" style="width:200px;"> General Matrix Multiplication.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEGEMM
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16<td>BFLOAT16
- </table>
-<tr>
- <td>CLGEMM
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>F16<td>F16<td>F16<td>F16
- </table>
-<tr>
- <td rowspan="1">GEMMConv2D
- <td rowspan="1" style="width:200px;"> General Matrix Multiplication.
- <td rowspan="1">
- <ul>
- <li>ANEURALNETWORKS_CONV_2D
- </ul>
- <td>NEGEMMConv2d
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16<td>BFLOAT16
- </table>
-<tr>
- <td rowspan="2">GEMMConvolutionLayer
- <td rowspan="2" style="width:200px;"> General Matrix Multiplication.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_CONV_2D
- </ul>
- <td>NEGEMMConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>BFLOAT16<td>BFLOAT16<td>BFLOAT16<td>BFLOAT16
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td>CLGEMMConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="1">GEMMDeconvolutionLayer
- <td rowspan="1" style="width:200px;"> General Matrix Multiplication.
- <td rowspan="1">
- <ul>
- <li>ANEURALNETWORKS_TRANSPOSE_CONV_2D
- </ul>
- <td>CLGEMMDeconvolutionLayer
- <td>
- <ul>
- <li>NHWC
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="2">GEMMLowpMatrixMultiplyCore
- <td rowspan="2" style="width:200px;"> General Matrix Multiplication.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEGEMMLowpMatrixMultiplyCore
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>S32
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>S32
- <tr><td>QASYMM8<td>QSYMM8<td>S32<td>S32
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>S32
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>S32
- <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>S32
- </table>
-<tr>
- <td>CLGEMMLowpMatrixMultiplyCore
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QSYMM8<td>S32<td>QASYMM8
- <tr><td>QASYMM8<td>QASYMM8<td>S32<td>S32
- <tr><td>QASYMM8<td>QSYMM8_PER_CHANNEL<td>S32<td>S32
- <tr><td>QASYMM8<td>QSYMM8<td>S32<td>S32
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>QASYMM8_SIGNED
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>S32<td>S32
- <tr><td>QASYMM8_SIGNED<td>QSYMM8_PER_CHANNEL<td>S32<td>S32
- <tr><td>QASYMM8_SIGNED<td>QSYMM8<td>S32<td>S32
- </table>
-<tr>
- <td rowspan="2">GEMMLowpOutputStage
- <td rowspan="2" style="width:200px;"> General Matrix Multiplication.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEGEMMLowpOutputStage
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>S32<td>S32<td>QASYMM8
- <tr><td>S32<td>S32<td>QASYMM8_SIGNED
- <tr><td>S32<td>S32<td>QSYMM16
- </table>
-<tr>
- <td>CLGEMMLowpOutputStage
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>S32<td>S32<td>QASYMM8
- <tr><td>S32<td>S32<td>QASYMM8_SIGNED
- <tr><td>S32<td>S32<td>QSYMM16
- </table>
-<tr>
- <td rowspan="2">GenerateProposalsLayer
- <td rowspan="2" style="width:200px;"> Function to generate proposals for a RPN (Region Proposal Network).
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_GENERATE_PROPOSALS
- </ul>
- <td>NEGenerateProposalsLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QSYMM8<td>QSYMM16<td>QASYMM8
- </table>
-<tr>
- <td>CLGenerateProposalsLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QSYMM8<td>QSYMM16<td>QASYMM8
- </table>
-<tr>
- <td rowspan="2">InstanceNormalizationLayer
- <td rowspan="2" style="width:200px;"> Function to perform a Instance normalization on a given axis.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_INSTANCE_NORMALIZATION
- </ul>
- <td>NEInstanceNormalizationLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLInstanceNormalizationLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">L2NormalizeLayer
- <td rowspan="2" style="width:200px;"> Function to perform a L2 normalization on a given axis.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_L2_NORMALIZATION
- </ul>
- <td>NEL2NormalizeLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLL2NormalizeLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="3">Logical
- <td rowspan="3" style="width:200px;"> Function to perform: - Logical AND - Logical OR - Logical NOT
- <td rowspan="3">
- <ul>
- <li>n/a
- </ul>
- <td>NELogicalAnd
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>U8<td>U8<td>U8
- </table>
-<tr>
- <td>NELogicalOr
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>U8<td>U8<td>U8
- </table>
-<tr>
- <td>NELogicalNot
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td rowspan="1">LogicalAnd
- <td rowspan="1" style="width:200px;"> Function to perform Logical AND.
- <td rowspan="1">
- <ul>
- <li>n/a
- </ul>
- <td>CLLogicalAnd
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>U8<td>U8<td>U8
- </table>
-<tr>
- <td rowspan="1">LogicalOr
- <td rowspan="1" style="width:200px;"> Function to perform Logical OR.
- <td rowspan="1">
- <ul>
- <li>n/a
- </ul>
- <td>CLLogicalOr
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>U8<td>U8<td>U8
- </table>
-<tr>
- <td rowspan="1">LogicalNot
- <td rowspan="1" style="width:200px;"> Function to perform Logical NOT.
- <td rowspan="1">
- <ul>
- <li>n/a
- </ul>
- <td>CLLogicalNot
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>U8<td>U8
- </table>
-<tr>
- <td rowspan="2">LSTMLayer
- <td rowspan="2" style="width:200px;"> Function to perform a single time step in a Long Short-Term Memory (LSTM) layer.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_LSTM
- </ul>
- <td>NELSTMLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0 - src13<th>dst0 - dst3
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLLSTMLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0 - src13<th>dst0 - dst3
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">LSTMLayerQuantized
- <td rowspan="2" style="width:200px;"> Function to perform quantized LSTM (Long Short-Term Memory)
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_QUANTIZED_LSTM
- <li>ANEURALNETWORKS_QUANTIZED_16BIT_LSTM
- </ul>
- <td>NELSTMLayerQuantized
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0 - src8<th>src9 - src12<th>src13<th>src14<th>dst0<th>dst1
- <tr><td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8<td>QSYMM16<td>QASYMM8
- </table>
-<tr>
- <td>CLLSTMLayerQuantized
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0 - src8<th>src9 - src12<th>src13<th>src14<th>dst0<th>dst1
- <tr><td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8<td>QSYMM16<td>QASYMM8
- </table>
-<tr>
- <td rowspan="2">MaxUnpoolingLayer
- <td rowspan="2" style="width:200px;"> Function to perform MaxUnpooling.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEMaxUnpoolingLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLMaxUnpoolingLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">MeanStdDevNormalizationLayer
- <td rowspan="2" style="width:200px;"> Function to execute mean and standard deviation normalization.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEMeanStdDevNormalizationLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td>CLMeanStdDevNormalizationLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td rowspan="2">NormalizationLayer
- <td rowspan="2" style="width:200px;"> Function to compute normalization layer.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_LOCAL_RESPONSE_NORMALIZATION
- </ul>
- <td>NENormalizationLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td>CLNormalizationLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F32<td>F32
- <tr><td>F16<td>F16
- </table>
-<tr>
- <td rowspan="2">PadLayer
- <td rowspan="2" style="width:200px;"> Function to pad a tensor.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_PAD
- <li>ANEURALNETWORKS_PAD_V2
- </ul>
- <td>NEPadLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLPadLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">Permute
- <td rowspan="2" style="width:200px;"> Function to transpose an ND tensor.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_TRANSPOSE
- </ul>
- <td>NEPermute
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLPermute
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">PixelWiseMultiplication
- <td rowspan="2" style="width:200px;"> Function to performe a multiplication.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_MUL
- </ul>
- <td>NEPixelWiseMultiplication
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
- <tr><td>QSYMM16<td>QSYMM16<td>S32
- <tr><td>U8<td>U8<td>U8
- <tr><td>U8<td>U8<td>S16
- <tr><td>U8<td>S16<td>S16
- <tr><td>S16<td>U8<td>S16
- <tr><td>S16<td>S16<td>S16
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>S32<td>F32
- </table>
-<tr>
- <td>CLPixelWiseMultiplication
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>QASYMM8<td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>QSYMM16<td>QSYMM16<td>QASYMM16
- <tr><td>QSYMM16<td>QSYMM16<td>S32
- <tr><td>U8<td>U8<td>U8
- <tr><td>U8<td>U8<td>S16
- <tr><td>U8<td>S16<td>S16
- <tr><td>S16<td>U8<td>S16
- <tr><td>S16<td>S16<td>S16
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>S32<td>F32
- </table>
-<tr>
- <td rowspan="2">PoolingLayer
- <td rowspan="2" style="width:200px;"> Function to performe pooling with the specified pooling operation.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_AVERAGE_POOL_2D
- <li>ANEURALNETWORKS_L2_POOL_2D
- <li>ANEURALNETWORKS_MAX_POOL_2D
- </ul>
- <td>NEPoolingLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLPoolingLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">PReluLayer
- <td rowspan="2" style="width:200px;"> Function to compute the activation layer with the PRELU activation function.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_PRELU
- </ul>
- <td>NEPReluLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLPReluLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">PriorBoxLayer
- <td rowspan="2" style="width:200px;"> Function to compute prior boxes and clip.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEPriorBoxLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLPriorBoxLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F32<td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">QLSTMLayer
- <td rowspan="2" style="width:200px;"> Function to perform quantized LSTM (Long Short-Term Memory).
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_QUANTIZED_LSTM
- <li>ANEURALNETWORKS_QUANTIZED_16BIT_LSTM
- </ul>
- <td>NEQLSTMLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1 - src6<th>src7 -src9<th>src10<th>src11<th>dst0<th>dst1 - dst2
- <tr><td>QASYMM8_SIGNED<td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8_SIGNED<td>QSYMM16<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td>CLQLSTMLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1 - src6<th>src7 -src9<th>src10<th>src11<th>dst0<th>dst1 - dst2
- <tr><td>QASYMM8_SIGNED<td>QASYMM8<td>S32<td>QSYMM16<td>QASYMM8_SIGNED<td>QSYMM16<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="2">QuantizationLayer
- <td rowspan="2" style="width:200px;"> Function to perform quantization layer
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_QUANTIZE
- </ul>
- <td>NEQuantizationLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
- <tr><td>QASYMM8_SIGNED<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
- <tr><td>F16<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
- <tr><td>F32<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
- </table>
-<tr>
- <td>CLQuantizationLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
- <tr><td>QASYMM8_SIGNED<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
- <tr><td>F16<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
- <tr><td>F32<td>QASYMM8, QASYMM8_SIGNED, QASYMM16
- </table>
-<tr>
- <td rowspan="2">Range
- <td rowspan="2" style="width:200px;"> Function to generates a sequence of numbers starting from START and extends by increments of 'STEP' up to but not including 'END'.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NERange
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>dst
- <tr><td>U8
- <tr><td>S8
- <tr><td>U16
- <tr><td>S16
- <tr><td>U32
- <tr><td>S32
- <tr><td>F16
- <tr><td>F32
- </table>
-<tr>
- <td>CLRange
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>dst
- <tr><td>U8
- <tr><td>S8
- <tr><td>QASYMM8
- <tr><td>U16
- <tr><td>S16
- <tr><td>U32
- <tr><td>S32
- <tr><td>F16
- <tr><td>F32
- </table>
-<tr>
- <td rowspan="2">ReduceMean
- <td rowspan="2" style="width:200px;"> Function to performe reduce mean operation.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_MEAN
- </ul>
- <td>NEReduceMean
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLReduceMean
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">ReductionOperation
- <td rowspan="2" style="width:200px;"> Function to performe reduce with the following operations - ARG_IDX_MAX: Index of the max value - ARG_IDX_MIN: Index of the min value - MEAN_SUM: Mean of sum - PROD: Product - SUM_SQUARE: Sum of squares - SUM: Sum - MIN: Min - MAX: Max
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_REDUCE_ALL
- <li>ANEURALNETWORKS_REDUCE_ANY
- <li>ANEURALNETWORKS_REDUCE_MAX
- <li>ANEURALNETWORKS_REDUCE_MIN
- <li>ANEURALNETWORKS_REDUCE_PROD
- <li>ANEURALNETWORKS_REDUCE_SUM
- </ul>
- <td>NEReductionOperation
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- <tr><td>S32<td>S32
- </table>
-<tr>
- <td>CLReductionOperation
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- <tr><td>S32<td>S32
- </table>
-<tr>
- <td rowspan="2">ReorgLayer
- <td rowspan="2" style="width:200px;"> Performs a reorganization layer of input tensor to the output tensor.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEReorgLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLReorgLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">ReshapeLayer
- <td rowspan="2" style="width:200px;"> Function to reshape a tensor.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_RESHAPE
- <li>ANEURALNETWORKS_SQUEEZE
- </ul>
- <td>NEReshapeLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLReshapeLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">Reverse
- <td rowspan="2" style="width:200px;"> Function to reverse tensor according to axis.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEReverse
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>All<td>U32<td>All
- </table>
-<tr>
- <td>CLReverse
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>All<td>U32<td>All
- </table>
-<tr>
- <td rowspan="2">RNNLayer
- <td rowspan="2" style="width:200px;"> Function to perform recurrent neural network layer.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_RNN
- </ul>
- <td>NERNNLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>src3<th>dst0<th>dst1
- <tr><td>F16<td>F16<td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLRNNLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>src3<th>dst0<th>dst1
- <tr><td>F16<td>F16<td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32<td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">ROIAlignLayer
- <td rowspan="2" style="width:200px;"> Function to perform ROI alignment.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_ROI_ALIGN
- </ul>
- <td>NEROIAlignLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM16<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM16<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td>CLROIAlignLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32
- <tr><td>QASYMM8<td>QASYMM16<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM16<td>QASYMM8_SIGNED
- </table>
-<tr>
- <td rowspan="2">ROIPoolingLayer
- <td rowspan="2" style="width:200px;"> Function to perform ROI pooling.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_ROI_POOLING
- </ul>
- <td>NEROIPoolingLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F32<td>U16<td>F32
- <tr><td>QASYMM8<td>U16<td>QASYMM8
- </table>
-<tr>
- <td>CLROIPoolingLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>dst
- <tr><td>F16<td>U16<td>F16
- <tr><td>F32<td>U16<td>F32
- <tr><td>QASYMM8<td>U16<td>QASYMM8
- </table>
-<tr>
- <td rowspan="2">Scale
- <td rowspan="2" style="width:200px;"> Function to perform resize a tensor using to interpolate: - Bilinear - Nearest neighbor
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_RESIZE_BILINEAR
- <li>ANEURALNETWORKS_RESIZE_NEAREST_NEIGHBOR
- </ul>
- <td>NEScale
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- <tr><td>U8<td>U8
- <tr><td>S16<td>S16
- </table>
-<tr>
- <td>CLScale
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- <tr><td>U8<td>U8
- <tr><td>S16<td>S16
- </table>
-<tr>
- <td rowspan="2">Select
- <td rowspan="2" style="width:200px;"> Function to select values from 2 tensors depending on an input tensor of booleans.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_SELECT
- </ul>
- <td>NESelect
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>U8<td>All<td>All<td>All
- </table>
-<tr>
- <td>CLSelect
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>U8<td>All<td>All<td>All
- </table>
-<tr>
- <td rowspan="2">Slice
- <td rowspan="2" style="width:200px;"> Function to perform tensor slicing.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_SLICE
- </ul>
- <td>NESlice
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLSlice
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">SoftmaxLayer
- <td rowspan="2" style="width:200px;"> Function to compute a SoftmaxLayer and a Log SoftmaxLayer.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_LOG_SOFTMAX
- <li>ANEURALNETWORKS_SOFTMAX
- </ul>
- <td>NESoftmaxLayerGeneric
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td>CLSoftmaxLayerGeneric
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>QASYMM8<td>QASYMM8
- <tr><td>QASYMM8_SIGNED<td>QASYMM8_SIGNED
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-<tr>
- <td rowspan="2">SpaceToBatchLayer
- <td rowspan="2" style="width:200px;"> Function to divide a tensor spatially.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_SPACE_TO_BATCH_ND
- </ul>
- <td>NESpaceToBatchLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>All<td>S32<td>S32<td>All
- </table>
-<tr>
- <td>CLSpaceToBatchLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>All<td>S32<td>S32<td>All
- </table>
-<tr>
- <td rowspan="2">SpaceToDepthLayer
- <td rowspan="2" style="width:200px;"> Function to rearrange blocks of spatial data into depth.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_SPACE_TO_DEPTH
- </ul>
- <td>NESpaceToDepthLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLSpaceToDepthLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">Split
- <td rowspan="2" style="width:200px;"> Function to split a tensor along a given axis.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_SPLIT
- </ul>
- <td>NESplit
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLSplit
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">StackLayer
- <td rowspan="2" style="width:200px;"> Function to stack tensors along an axis.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEStackLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLStackLayer
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">StridedSlice
- <td rowspan="2" style="width:200px;"> Function to extract a strided slice of a tensor.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_STRIDED_SLICE
- </ul>
- <td>NEStridedSlice
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLStridedSlice
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">Tile
- <td rowspan="2" style="width:200px;"> Function to construct a tensor by tiling a given tensor.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_TILE
- </ul>
- <td>NETile
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLTile
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">Transpose
- <td rowspan="2" style="width:200px;"> Function to transpose a 2D tensor.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_TRANSPOSE
- </ul>
- <td>NETranspose
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLTranspose
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">Unstack
- <td rowspan="2" style="width:200px;"> Function to unpack a rank-R tensor into rank-(R-1) tensors.
- <td rowspan="2">
- <ul>
- <li>n/a
- </ul>
- <td>NEUnstack
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td>CLUnstack
- <td>
- <ul>
- <li>All
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>All<td>All
- </table>
-<tr>
- <td rowspan="2">WinogradConvolutionLayer
- <td rowspan="2" style="width:200px;"> Function to do Winograd Convolution.
- <td rowspan="2">
- <ul>
- <li>ANEURALNETWORKS_CONV_2D
- </ul>
- <td>NEWinogradConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- </table>
-<tr>
- <td>CLWinogradConvolutionLayer
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src0<th>src1<th>src2<th>dst
- <tr><td>F16<td>F16<td>F16<td>F16
- <tr><td>F32<td>F32<td>F32<td>F32
- </table>
-<tr>
- <td rowspan="1">WinogradInputTransform
- <td rowspan="1" style="width:200px;"> Function to.
- <td rowspan="1">
- <ul>
- <li>n/a
- </ul>
- <td>CLWinogradInputTransform
- <td>
- <ul>
- <li>NHWC
- <li>NCHW
- </ul>
- <td>
- <table>
- <tr><th>src<th>dst
- <tr><td>F16<td>F16
- <tr><td>F32<td>F32
- </table>
-</table>
-
-*/
-} // namespace \ No newline at end of file