aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--Android.mk70
-rw-r--r--CMakeLists.txt140
-rw-r--r--src/armnn/backends/ClLayerSupport.cpp18
-rw-r--r--src/armnn/backends/ClWorkloadFactory.cpp36
-rw-r--r--src/armnn/backends/ClWorkloads.hpp36
-rw-r--r--src/armnn/backends/ClWorkloads/ClActivationFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClActivationFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/ClWorkloads/ClActivationFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClActivationFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/ClWorkloads/ClAdditionFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClAdditionFloat32Workload.cpp)6
-rw-r--r--src/armnn/backends/ClWorkloads/ClAdditionFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClAdditionFloat32Workload.hpp)2
-rw-r--r--src/armnn/backends/ClWorkloads/ClBatchNormalizationFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClBatchNormalizationFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/ClWorkloads/ClBatchNormalizationFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClBatchNormalizationFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/ClWorkloads/ClConstantFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClConstantFloat32Workload.cpp)6
-rw-r--r--src/armnn/backends/ClWorkloads/ClConstantFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClConstantFloat32Workload.hpp)2
-rw-r--r--src/armnn/backends/ClWorkloads/ClConvolution2dFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClConvolution2dFloat32Workload.cpp)8
-rw-r--r--src/armnn/backends/ClWorkloads/ClConvolution2dFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClConvolution2dFloat32Workload.hpp)6
-rw-r--r--src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.cpp)8
-rw-r--r--src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.hpp)8
-rw-r--r--src/armnn/backends/ClWorkloads/ClFloorFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClFloorFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/ClWorkloads/ClFloorFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClFloorFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/ClWorkloads/ClFullyConnectedFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClFullyConnectedFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/ClWorkloads/ClFullyConnectedFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClFullyConnectedFloat32Workload.hpp)8
-rw-r--r--src/armnn/backends/ClWorkloads/ClL2NormalizationFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClL2NormalizationFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/ClWorkloads/ClL2NormalizationFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClL2NormalizationFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.hpp67
-rw-r--r--src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.cpp)48
-rw-r--r--src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.hpp67
-rw-r--r--src/armnn/backends/ClWorkloads/ClMergerFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClMergerFloat32Workload.cpp)6
-rw-r--r--src/armnn/backends/ClWorkloads/ClMergerFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClMergerFloat32Workload.hpp)2
-rw-r--r--src/armnn/backends/ClWorkloads/ClMultiplicationFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClMultiplicationFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/ClWorkloads/ClMultiplicationFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClMultiplicationFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/ClWorkloads/ClNormalizationFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClNormalizationFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/ClWorkloads/ClNormalizationFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClNormalizationFloat32Workload.hpp)7
-rw-r--r--src/armnn/backends/ClWorkloads/ClPooling2dFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClPooling2dFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/ClWorkloads/ClPooling2dFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClPooling2dFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/ClWorkloads/ClReshapeFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClReshapeFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/ClWorkloads/ClReshapeFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClReshapeFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/ClWorkloads/ClResizeBilinearFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClResizeBilinearFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/ClWorkloads/ClResizeBilinearFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClResizeBilinearFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/ClWorkloads/ClSoftmaxFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClSoftmaxFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/ClWorkloads/ClSoftmaxFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClSoftmaxFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/ClWorkloads/ClSplitterFloatWorkload.cpp (renamed from src/armnn/backends/ClWorkloads/ClSplitterFloat32Workload.cpp)6
-rw-r--r--src/armnn/backends/ClWorkloads/ClSplitterFloatWorkload.hpp (renamed from src/armnn/backends/ClWorkloads/ClSplitterFloat32Workload.hpp)2
-rw-r--r--src/armnn/backends/NeonLayerSupport.cpp14
-rw-r--r--src/armnn/backends/NeonWorkloadFactory.cpp34
-rw-r--r--src/armnn/backends/NeonWorkloads.hpp34
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonActivationFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonActivationFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonActivationFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonActivationFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonAdditionFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonAdditionFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonAdditionFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonAdditionFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.hpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonConstantFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonConstantFloat32Workload.cpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonConstantFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonConstantFloat32Workload.hpp)2
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonConvolution2dFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonConvolution2dFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonConvolution2dFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonConvolution2dFloat32Workload.hpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.hpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.cpp30
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonFloorFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.hpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.hpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonLstmFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonLstmFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonLstmFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonLstmFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonMergerFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonMergerFloat32Workload.cpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonMergerFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonMergerFloat32Workload.hpp)2
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonMultiplicationFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonMultiplicationFloat32Workload.cpp)12
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonMultiplicationFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonMultiplicationFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonNormalizationFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonNormalizationFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonNormalizationFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonNormalizationFloat32Workload.hpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.cpp27
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.cpp27
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.hpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.cpp32
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonReshapeFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonFloorFloat32Workload.cpp)18
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonReshapeFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.hpp)4
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonSoftmaxFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonSoftmaxFloat32Workload.cpp)10
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonSoftmaxFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonSoftmaxFloat32Workload.hpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonSplitterFloatWorkload.cpp (renamed from src/armnn/backends/NeonWorkloads/NeonSplitterFloat32Workload.cpp)6
-rw-r--r--src/armnn/backends/NeonWorkloads/NeonSplitterFloatWorkload.hpp (renamed from src/armnn/backends/NeonWorkloads/NeonSplitterFloat32Workload.hpp)2
-rw-r--r--src/armnn/backends/test/CreateWorkloadCl.cpp102
-rw-r--r--src/armnn/backends/test/CreateWorkloadNeon.cpp80
83 files changed, 665 insertions, 666 deletions
diff --git a/Android.mk b/Android.mk
index c81b99737d..c712ced385 100644
--- a/Android.mk
+++ b/Android.mk
@@ -43,76 +43,76 @@ LOCAL_SRC_FILES := \
src/armnnUtils/Logging.cpp \
src/armnnUtils/Permute.cpp \
src/armnn/backends/ArmComputeTensorUtils.cpp \
- src/armnn/backends/ClWorkloads/ClActivationFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClActivationFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClActivationUint8Workload.cpp \
src/armnn/backends/ClWorkloads/ClAdditionBaseWorkload.cpp \
- src/armnn/backends/ClWorkloads/ClAdditionFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClAdditionFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClAdditionUint8Workload.cpp \
src/armnn/backends/ClWorkloads/ClBaseConstantWorkload.cpp \
- src/armnn/backends/ClWorkloads/ClBatchNormalizationFloat32Workload.cpp \
- src/armnn/backends/ClWorkloads/ClConstantFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClBatchNormalizationFloatWorkload.cpp \
+ src/armnn/backends/ClWorkloads/ClConstantFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClConstantUint8Workload.cpp \
src/armnn/backends/ClWorkloads/ClConvertFp16ToFp32Workload.cpp \
src/armnn/backends/ClWorkloads/ClConvertFp32ToFp16Workload.cpp \
src/armnn/backends/ClWorkloads/ClConvolution2dBaseWorkload.cpp \
- src/armnn/backends/ClWorkloads/ClConvolution2dFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClConvolution2dFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClConvolution2dUint8Workload.cpp \
src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionBaseWorkload.cpp \
- src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionUint8Workload.cpp \
src/armnn/backends/ClWorkloads/ClDivisionFloatWorkload.cpp \
- src/armnn/backends/ClWorkloads/ClFloorFloat32Workload.cpp \
- src/armnn/backends/ClWorkloads/ClFullyConnectedFloat32Workload.cpp \
- src/armnn/backends/ClWorkloads/ClL2NormalizationFloat32Workload.cpp \
- src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.cpp \
- src/armnn/backends/ClWorkloads/ClMergerFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClFloorFloatWorkload.cpp \
+ src/armnn/backends/ClWorkloads/ClFullyConnectedFloatWorkload.cpp \
+ src/armnn/backends/ClWorkloads/ClL2NormalizationFloatWorkload.cpp \
+ src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.cpp \
+ src/armnn/backends/ClWorkloads/ClMergerFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClMergerUint8Workload.cpp \
- src/armnn/backends/ClWorkloads/ClMultiplicationFloat32Workload.cpp \
- src/armnn/backends/ClWorkloads/ClNormalizationFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClMultiplicationFloatWorkload.cpp \
+ src/armnn/backends/ClWorkloads/ClNormalizationFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClPermuteWorkload.cpp \
src/armnn/backends/ClWorkloads/ClPooling2dBaseWorkload.cpp \
- src/armnn/backends/ClWorkloads/ClPooling2dFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClPooling2dFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClPooling2dUint8Workload.cpp \
- src/armnn/backends/ClWorkloads/ClReshapeFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClReshapeFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClReshapeUint8Workload.cpp \
- src/armnn/backends/ClWorkloads/ClResizeBilinearFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClResizeBilinearFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClSoftmaxBaseWorkload.cpp \
- src/armnn/backends/ClWorkloads/ClSoftmaxFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClSoftmaxFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClSoftmaxUint8Workload.cpp \
- src/armnn/backends/ClWorkloads/ClSplitterFloat32Workload.cpp \
+ src/armnn/backends/ClWorkloads/ClSplitterFloatWorkload.cpp \
src/armnn/backends/ClWorkloads/ClSplitterUint8Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonActivationFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonActivationFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonActivationUint8Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonAdditionFloat32Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonConstantFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonAdditionFloatWorkload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonConstantFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonConstantUint8Workload.cpp \
src/armnn/backends/NeonWorkloads/NeonConvertFp16ToFp32Workload.cpp \
src/armnn/backends/NeonWorkloads/NeonConvertFp32ToFp16Workload.cpp \
src/armnn/backends/NeonWorkloads/NeonConvolution2dBaseWorkload.cpp \
- src/armnn/backends/NeonWorkloads/NeonConvolution2dFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonConvolution2dFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonConvolution2dUint8Workload.cpp \
src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionBaseWorkload.cpp \
- src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionUint8Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonFloorFloat32Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonLstmFloat32Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonMergerFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonLstmFloatWorkload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonMergerFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonMergerUint8Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonMultiplicationFloat32Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonNormalizationFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonMultiplicationFloatWorkload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonNormalizationFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonPermuteWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonPooling2dBaseWorkload.cpp \
- src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonPooling2dUint8Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonReshapeFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonReshapeUint8Workload.cpp \
src/armnn/backends/NeonWorkloads/NeonSoftmaxBaseWorkload.cpp \
- src/armnn/backends/NeonWorkloads/NeonSoftmaxFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonSoftmaxFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonSoftmaxUint8Workload.cpp \
- src/armnn/backends/NeonWorkloads/NeonSplitterFloat32Workload.cpp \
+ src/armnn/backends/NeonWorkloads/NeonSplitterFloatWorkload.cpp \
src/armnn/backends/NeonWorkloads/NeonSplitterUint8Workload.cpp \
src/armnn/backends/ClWorkloadFactory.cpp \
src/armnn/backends/ClContextControl.cpp \
diff --git a/CMakeLists.txt b/CMakeLists.txt
index 97b50df623..09761af4b1 100644
--- a/CMakeLists.txt
+++ b/CMakeLists.txt
@@ -395,19 +395,19 @@ list(APPEND armnn_sources
if(ARMCOMPUTENEON)
# Additionally include source files for ARM Compute NEON backend
list(APPEND armnn_sources
- src/armnn/backends/NeonWorkloads/NeonActivationFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonActivationFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonActivationFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonActivationFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonActivationUint8Workload.cpp
src/armnn/backends/NeonWorkloads/NeonActivationUint8Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonAdditionFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonAdditionFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonAdditionFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonAdditionFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonBaseConstantWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonBaseMergerWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonBaseSplitterWorkload.hpp
- src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonConstantFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonConstantFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.hpp
+ src/armnn/backends/NeonWorkloads/NeonConstantFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonConstantFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonConstantUint8Workload.cpp
src/armnn/backends/NeonWorkloads/NeonConstantUint8Workload.hpp
src/armnn/backends/NeonWorkloads/NeonConvertFp16ToFp32Workload.cpp
@@ -416,52 +416,52 @@ if(ARMCOMPUTENEON)
src/armnn/backends/NeonWorkloads/NeonConvertFp32ToFp16Workload.hpp
src/armnn/backends/NeonWorkloads/NeonConvolution2dBaseWorkload.cpp
src/armnn/backends/NeonWorkloads/NeonConvolution2dBaseWorkload.hpp
- src/armnn/backends/NeonWorkloads/NeonConvolution2dFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonConvolution2dFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonConvolution2dFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonConvolution2dFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonConvolution2dUint8Workload.cpp
src/armnn/backends/NeonWorkloads/NeonConvolution2dUint8Workload.hpp
src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionBaseWorkload.cpp
src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionBaseWorkload.hpp
- src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionUint8Workload.cpp
src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionUint8Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonFloorFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonFloorFloat32Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonLstmFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonLstmFloat32Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonMergerFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonMergerFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.hpp
+ src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.hpp
+ src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.hpp
+ src/armnn/backends/NeonWorkloads/NeonLstmFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonLstmFloatWorkload.hpp
+ src/armnn/backends/NeonWorkloads/NeonMergerFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonMergerFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonMergerUint8Workload.cpp
src/armnn/backends/NeonWorkloads/NeonMergerUint8Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonMultiplicationFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonMultiplicationFloat32Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonNormalizationFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonNormalizationFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonMultiplicationFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonMultiplicationFloatWorkload.hpp
+ src/armnn/backends/NeonWorkloads/NeonNormalizationFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonNormalizationFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonPermuteWorkload.cpp
src/armnn/backends/NeonWorkloads/NeonPermuteWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonPooling2dBaseWorkload.cpp
src/armnn/backends/NeonWorkloads/NeonPooling2dBaseWorkload.hpp
- src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonPooling2dUint8Workload.cpp
src/armnn/backends/NeonWorkloads/NeonPooling2dUint8Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonReshapeFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonReshapeFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonReshapeUint8Workload.cpp
src/armnn/backends/NeonWorkloads/NeonReshapeUint8Workload.hpp
src/armnn/backends/NeonWorkloads/NeonSoftmaxBaseWorkload.cpp
src/armnn/backends/NeonWorkloads/NeonSoftmaxBaseWorkload.hpp
- src/armnn/backends/NeonWorkloads/NeonSoftmaxFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonSoftmaxFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonSoftmaxFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonSoftmaxFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonSoftmaxUint8Workload.cpp
src/armnn/backends/NeonWorkloads/NeonSoftmaxUint8Workload.hpp
- src/armnn/backends/NeonWorkloads/NeonSplitterFloat32Workload.cpp
- src/armnn/backends/NeonWorkloads/NeonSplitterFloat32Workload.hpp
+ src/armnn/backends/NeonWorkloads/NeonSplitterFloatWorkload.cpp
+ src/armnn/backends/NeonWorkloads/NeonSplitterFloatWorkload.hpp
src/armnn/backends/NeonWorkloads/NeonSplitterUint8Workload.cpp
src/armnn/backends/NeonWorkloads/NeonSplitterUint8Workload.hpp
src/armnn/backends/NeonWorkloadUtils.cpp
@@ -475,8 +475,8 @@ endif()
if(ARMCOMPUTECL)
# Additionally include source files for ARM Compute OpenCL backend
list(APPEND armnn_sources
- src/armnn/backends/ClWorkloads/ClActivationFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClActivationFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClActivationFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClActivationFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClActivationUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClActivationUint8Workload.hpp
src/armnn/backends/ClWorkloads/ClAdditionBaseWorkload.cpp
@@ -485,71 +485,71 @@ if(ARMCOMPUTECL)
src/armnn/backends/ClWorkloads/ClConvertFp16ToFp32Workload.hpp
src/armnn/backends/ClWorkloads/ClConvertFp32ToFp16Workload.cpp
src/armnn/backends/ClWorkloads/ClConvertFp32ToFp16Workload.hpp
- src/armnn/backends/ClWorkloads/ClAdditionFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClAdditionFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClAdditionFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClAdditionFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClAdditionUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClAdditionUint8Workload.hpp
src/armnn/backends/ClWorkloads/ClBaseConstantWorkload.cpp
src/armnn/backends/ClWorkloads/ClBaseConstantWorkload.hpp
src/armnn/backends/ClWorkloads/ClBaseMergerWorkload.hpp
- src/armnn/backends/ClWorkloads/ClBatchNormalizationFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClBatchNormalizationFloat32Workload.hpp
- src/armnn/backends/ClWorkloads/ClConstantFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClConstantFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClBatchNormalizationFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClBatchNormalizationFloatWorkload.hpp
+ src/armnn/backends/ClWorkloads/ClConstantFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClConstantFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClConstantUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClConstantUint8Workload.hpp
src/armnn/backends/ClWorkloads/ClConvolution2dBaseWorkload.cpp
src/armnn/backends/ClWorkloads/ClConvolution2dBaseWorkload.hpp
- src/armnn/backends/ClWorkloads/ClConvolution2dFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClConvolution2dFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClConvolution2dFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClConvolution2dFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClConvolution2dUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClConvolution2dUint8Workload.hpp
src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionBaseWorkload.cpp
src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionBaseWorkload.hpp
- src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionUint8Workload.hpp
- src/armnn/backends/ClWorkloads/ClFloorFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClFloorFloat32Workload.hpp
- src/armnn/backends/ClWorkloads/ClFullyConnectedFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClFullyConnectedFloat32Workload.hpp
- src/armnn/backends/ClWorkloads/ClL2NormalizationFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClL2NormalizationFloat32Workload.hpp
- src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.hpp
- src/armnn/backends/ClWorkloads/ClMergerFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClMergerFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClFloorFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClFloorFloatWorkload.hpp
+ src/armnn/backends/ClWorkloads/ClFullyConnectedFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClFullyConnectedFloatWorkload.hpp
+ src/armnn/backends/ClWorkloads/ClL2NormalizationFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClL2NormalizationFloatWorkload.hpp
+ src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.hpp
+ src/armnn/backends/ClWorkloads/ClMergerFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClMergerFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClMergerUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClMergerUint8Workload.hpp
- src/armnn/backends/ClWorkloads/ClMultiplicationFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClMultiplicationFloat32Workload.hpp
- src/armnn/backends/ClWorkloads/ClNormalizationFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClNormalizationFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClMultiplicationFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClMultiplicationFloatWorkload.hpp
+ src/armnn/backends/ClWorkloads/ClNormalizationFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClNormalizationFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClPermuteWorkload.cpp
src/armnn/backends/ClWorkloads/ClPermuteWorkload.hpp
src/armnn/backends/ClWorkloads/ClPooling2dBaseWorkload.cpp
src/armnn/backends/ClWorkloads/ClPooling2dBaseWorkload.hpp
- src/armnn/backends/ClWorkloads/ClPooling2dFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClPooling2dFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClPooling2dFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClPooling2dFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClPooling2dUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClPooling2dUint8Workload.hpp
- src/armnn/backends/ClWorkloads/ClReshapeFloat32Workload.cpp
+ src/armnn/backends/ClWorkloads/ClReshapeFloatWorkload.cpp
src/armnn/backends/ClWorkloads/ClDivisionFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClDivisionFloatWorkload.cpp
- src/armnn/backends/ClWorkloads/ClReshapeFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClReshapeFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClReshapeUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClReshapeUint8Workload.hpp
- src/armnn/backends/ClWorkloads/ClResizeBilinearFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClResizeBilinearFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClResizeBilinearFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClResizeBilinearFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClSoftmaxBaseWorkload.cpp
src/armnn/backends/ClWorkloads/ClSoftmaxBaseWorkload.hpp
- src/armnn/backends/ClWorkloads/ClSoftmaxFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClSoftmaxFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClSoftmaxFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClSoftmaxFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClSoftmaxUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClSoftmaxUint8Workload.hpp
- src/armnn/backends/ClWorkloads/ClSplitterFloat32Workload.cpp
- src/armnn/backends/ClWorkloads/ClSplitterFloat32Workload.hpp
+ src/armnn/backends/ClWorkloads/ClSplitterFloatWorkload.cpp
+ src/armnn/backends/ClWorkloads/ClSplitterFloatWorkload.hpp
src/armnn/backends/ClWorkloads/ClSplitterUint8Workload.cpp
src/armnn/backends/ClWorkloads/ClSplitterUint8Workload.hpp
src/armnn/backends/ClWorkloadUtils.hpp
diff --git a/src/armnn/backends/ClLayerSupport.cpp b/src/armnn/backends/ClLayerSupport.cpp
index 77e74f57c6..623bdcafc9 100644
--- a/src/armnn/backends/ClLayerSupport.cpp
+++ b/src/armnn/backends/ClLayerSupport.cpp
@@ -14,22 +14,22 @@
#include <boost/core/ignore_unused.hpp>
#ifdef ARMCOMPUTECL_ENABLED
-#include "ClWorkloads/ClAdditionFloat32Workload.hpp"
-#include "ClWorkloads/ClActivationFloat32Workload.hpp"
-#include "ClWorkloads/ClBatchNormalizationFloat32Workload.hpp"
+#include "ClWorkloads/ClAdditionFloatWorkload.hpp"
+#include "ClWorkloads/ClActivationFloatWorkload.hpp"
+#include "ClWorkloads/ClBatchNormalizationFloatWorkload.hpp"
#include "ClWorkloads/ClConvertFp16ToFp32Workload.hpp"
#include "ClWorkloads/ClConvertFp32ToFp16Workload.hpp"
#include "ClWorkloads/ClConvolution2dBaseWorkload.hpp"
#include "ClWorkloads/ClDepthwiseConvolutionBaseWorkload.hpp"
#include "ClWorkloads/ClDivisionFloatWorkload.hpp"
-#include "ClWorkloads/ClL2NormalizationFloat32Workload.hpp"
-#include "ClWorkloads/ClMultiplicationFloat32Workload.hpp"
-#include "ClWorkloads/ClFullyConnectedFloat32Workload.hpp"
+#include "ClWorkloads/ClL2NormalizationFloatWorkload.hpp"
+#include "ClWorkloads/ClMultiplicationFloatWorkload.hpp"
+#include "ClWorkloads/ClFullyConnectedFloatWorkload.hpp"
#include "ClWorkloads/ClPooling2dBaseWorkload.hpp"
#include "ClWorkloads/ClPermuteWorkload.hpp"
-#include "ClWorkloads/ClNormalizationFloat32Workload.hpp"
+#include "ClWorkloads/ClNormalizationFloatWorkload.hpp"
#include "ClWorkloads/ClSoftmaxBaseWorkload.hpp"
-#include "ClWorkloads/ClLstmFloat32Workload.hpp"
+#include "ClWorkloads/ClLstmFloatWorkload.hpp"
#endif
using namespace boost;
@@ -417,7 +417,7 @@ bool IsLstmSupportedCl(const TensorInfo& input, const TensorInfo& outputStateIn,
const TensorInfo* projectionBias, const TensorInfo* cellToForgetWeights,
const TensorInfo* cellToOutputWeights, std::string* reasonIfUnsupported)
{
- FORWARD_WORKLOAD_VALIDATE_FUNC(ClLstmFloat32WorkloadValidate, reasonIfUnsupported,
+ FORWARD_WORKLOAD_VALIDATE_FUNC(ClLstmFloatWorkloadValidate, reasonIfUnsupported,
input, outputStateIn, cellStateIn, scratchBuffer, outputStateOut, cellStateOut,
output, descriptor, inputToForgetWeights, inputToCellWeights,
inputToOutputWeights, recurrentToForgetWeights,
diff --git a/src/armnn/backends/ClWorkloadFactory.cpp b/src/armnn/backends/ClWorkloadFactory.cpp
index 77959d1e64..86624cea23 100644
--- a/src/armnn/backends/ClWorkloadFactory.cpp
+++ b/src/armnn/backends/ClWorkloadFactory.cpp
@@ -91,32 +91,32 @@ std::unique_ptr<IWorkload> ClWorkloadFactory::CreateOutput(const OutputQueueDesc
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClActivationFloat32Workload, ClActivationUint8Workload>(descriptor, info);
+ return MakeWorkload<ClActivationFloatWorkload, ClActivationUint8Workload>(descriptor, info);
}
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClSoftmaxFloat32Workload, ClSoftmaxUint8Workload>(descriptor, info,
+ return MakeWorkload<ClSoftmaxFloatWorkload, ClSoftmaxUint8Workload>(descriptor, info,
m_MemoryManager.GetIntraLayerManager());
}
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClSplitterFloat32Workload, ClSplitterUint8Workload>(descriptor, info);
+ return MakeWorkload<ClSplitterFloatWorkload, ClSplitterUint8Workload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateMerger(const MergerQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClMergerFloat32Workload, ClMergerUint8Workload>(descriptor, info);
+ return MakeWorkload<ClMergerFloatWorkload, ClMergerUint8Workload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateFullyConnected(
const FullyConnectedQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<ClFullyConnectedFloat32Workload, NullWorkload>(descriptor, info,
+ return MakeWorkload<ClFullyConnectedFloatWorkload, NullWorkload>(descriptor, info,
m_MemoryManager.GetIntraLayerManager());
}
@@ -129,38 +129,38 @@ std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreatePermute(const Permute
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClPooling2dFloat32Workload, ClPooling2dUint8Workload>(descriptor, info);
+ return MakeWorkload<ClPooling2dFloatWorkload, ClPooling2dUint8Workload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateConvolution2d(const Convolution2dQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClConvolution2dFloat32Workload, ClConvolution2dUint8Workload>(descriptor, info,
+ return MakeWorkload<ClConvolution2dFloatWorkload, ClConvolution2dUint8Workload>(descriptor, info,
m_MemoryManager.GetIntraLayerManager());
}
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateDepthwiseConvolution2d(
const DepthwiseConvolution2dQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<ClDepthwiseConvolutionFloat32Workload, ClDepthwiseConvolutionUint8Workload>(descriptor, info);
+ return MakeWorkload<ClDepthwiseConvolutionFloatWorkload, ClDepthwiseConvolutionUint8Workload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateNormalization(const NormalizationQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClNormalizationFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<ClNormalizationFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClAdditionFloat32Workload, ClAdditionUint8Workload>(descriptor, info);
+ return MakeWorkload<ClAdditionFloatWorkload, ClAdditionUint8Workload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateMultiplication(
const MultiplicationQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<ClMultiplicationFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<ClMultiplicationFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateDivision(
@@ -172,7 +172,7 @@ std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateDivision(
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateBatchNormalization(
const BatchNormalizationQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<ClBatchNormalizationFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<ClBatchNormalizationFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor,
@@ -190,7 +190,7 @@ std::unique_ptr<armnn::IWorkload> ClWorkloadFactory::CreateResizeBilinear(
const ResizeBilinearQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClResizeBilinearFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<ClResizeBilinearFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFakeQuantization(
@@ -203,31 +203,31 @@ std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFakeQuantization(
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClL2NormalizationFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<ClL2NormalizationFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClConstantFloat32Workload, ClConstantUint8Workload>(descriptor, info);
+ return MakeWorkload<ClConstantFloatWorkload, ClConstantUint8Workload>(descriptor, info);
}
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClReshapeFloat32Workload, ClReshapeUint8Workload>(descriptor, info);
+ return MakeWorkload<ClReshapeFloatWorkload, ClReshapeUint8Workload>(descriptor, info);
}
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClFloorFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<ClFloorFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<ClLstmFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<ClLstmFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateConvertFp16ToFp32(
diff --git a/src/armnn/backends/ClWorkloads.hpp b/src/armnn/backends/ClWorkloads.hpp
index 6eb8adabe4..001083019f 100644
--- a/src/armnn/backends/ClWorkloads.hpp
+++ b/src/armnn/backends/ClWorkloads.hpp
@@ -4,37 +4,37 @@
//
#pragma once
-#include "backends/ClWorkloads/ClActivationFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClActivationFloatWorkload.hpp"
#include "backends/ClWorkloads/ClActivationUint8Workload.hpp"
-#include "backends/ClWorkloads/ClAdditionFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClAdditionFloatWorkload.hpp"
#include "backends/ClWorkloads/ClAdditionUint8Workload.hpp"
#include "backends/ClWorkloads/ClBaseConstantWorkload.hpp"
#include "backends/ClWorkloads/ClBaseMergerWorkload.hpp"
-#include "backends/ClWorkloads/ClBatchNormalizationFloat32Workload.hpp"
-#include "backends/ClWorkloads/ClConstantFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClBatchNormalizationFloatWorkload.hpp"
+#include "backends/ClWorkloads/ClConstantFloatWorkload.hpp"
#include "backends/ClWorkloads/ClConstantUint8Workload.hpp"
-#include "backends/ClWorkloads/ClConvolution2dFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClConvolution2dFloatWorkload.hpp"
#include "backends/ClWorkloads/ClConvolution2dUint8Workload.hpp"
-#include "backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.hpp"
#include "backends/ClWorkloads/ClDepthwiseConvolutionUint8Workload.hpp"
#include "backends/ClWorkloads/ClDivisionFloatWorkload.hpp"
-#include "backends/ClWorkloads/ClFloorFloat32Workload.hpp"
-#include "backends/ClWorkloads/ClFullyConnectedFloat32Workload.hpp"
-#include "backends/ClWorkloads/ClL2NormalizationFloat32Workload.hpp"
-#include "backends/ClWorkloads/ClLstmFloat32Workload.hpp"
-#include "backends/ClWorkloads/ClMergerFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClFloorFloatWorkload.hpp"
+#include "backends/ClWorkloads/ClFullyConnectedFloatWorkload.hpp"
+#include "backends/ClWorkloads/ClL2NormalizationFloatWorkload.hpp"
+#include "backends/ClWorkloads/ClLstmFloatWorkload.hpp"
+#include "backends/ClWorkloads/ClMergerFloatWorkload.hpp"
#include "backends/ClWorkloads/ClMergerUint8Workload.hpp"
-#include "backends/ClWorkloads/ClMultiplicationFloat32Workload.hpp"
-#include "backends/ClWorkloads/ClNormalizationFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClMultiplicationFloatWorkload.hpp"
+#include "backends/ClWorkloads/ClNormalizationFloatWorkload.hpp"
#include "backends/ClWorkloads/ClPermuteWorkload.hpp"
-#include "backends/ClWorkloads/ClPooling2dFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClPooling2dFloatWorkload.hpp"
#include "backends/ClWorkloads/ClPooling2dUint8Workload.hpp"
-#include "backends/ClWorkloads/ClReshapeFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClReshapeFloatWorkload.hpp"
#include "backends/ClWorkloads/ClReshapeUint8Workload.hpp"
-#include "backends/ClWorkloads/ClResizeBilinearFloat32Workload.hpp"
-#include "backends/ClWorkloads/ClSoftmaxFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClResizeBilinearFloatWorkload.hpp"
+#include "backends/ClWorkloads/ClSoftmaxFloatWorkload.hpp"
#include "backends/ClWorkloads/ClSoftmaxUint8Workload.hpp"
-#include "backends/ClWorkloads/ClSplitterFloat32Workload.hpp"
+#include "backends/ClWorkloads/ClSplitterFloatWorkload.hpp"
#include "backends/ClWorkloads/ClSplitterUint8Workload.hpp"
#include "backends/ClWorkloads/ClConvertFp16ToFp32Workload.hpp"
#include "backends/ClWorkloads/ClConvertFp32ToFp16Workload.hpp"
diff --git a/src/armnn/backends/ClWorkloads/ClActivationFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClActivationFloatWorkload.cpp
index f072549cbc..1e5cf0c722 100644
--- a/src/armnn/backends/ClWorkloads/ClActivationFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClActivationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClActivationFloat32Workload.hpp"
+#include "ClActivationFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/ArmComputeUtils.hpp"
@@ -31,11 +31,11 @@ arm_compute::Status ClActivationWorkloadValidate(const TensorInfo& input,
activationLayerInfo);
}
-ClActivationFloat32Workload::ClActivationFloat32Workload(const ActivationQueueDescriptor& descriptor,
+ClActivationFloatWorkload::ClActivationFloatWorkload(const ActivationQueueDescriptor& descriptor,
const WorkloadInfo& info)
: FloatWorkload<ActivationQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("ClActivationFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("ClActivationFloatWorkload", 1, 1);
const arm_compute::ActivationLayerInfo activationLayerInfo =
ConvertActivationDescriptorToAclActivationLayerInfo(m_Data.m_Parameters);
@@ -45,9 +45,9 @@ ClActivationFloat32Workload::ClActivationFloat32Workload(const ActivationQueueDe
m_ActivationLayer.configure(&input, &output, activationLayerInfo);
}
-void ClActivationFloat32Workload::Execute() const
+void ClActivationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClActivationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClActivationFloatWorkload_Execute");
m_ActivationLayer.run();
}
diff --git a/src/armnn/backends/ClWorkloads/ClActivationFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClActivationFloatWorkload.hpp
index 9fbfe95856..56367ce14f 100644
--- a/src/armnn/backends/ClWorkloads/ClActivationFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClActivationFloatWorkload.hpp
@@ -14,10 +14,10 @@ arm_compute::Status ClActivationWorkloadValidate(const TensorInfo& input,
const ActivationDescriptor& descriptor);
// Activation layer execution.
-class ClActivationFloat32Workload : public FloatWorkload<ActivationQueueDescriptor>
+class ClActivationFloatWorkload : public FloatWorkload<ActivationQueueDescriptor>
{
public:
- ClActivationFloat32Workload(const ActivationQueueDescriptor& descriptor, const WorkloadInfo& info);
+ ClActivationFloatWorkload(const ActivationQueueDescriptor& descriptor, const WorkloadInfo& info);
void Execute() const override;
private:
diff --git a/src/armnn/backends/ClWorkloads/ClAdditionFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClAdditionFloatWorkload.cpp
index b69593f5f5..089b84a33f 100644
--- a/src/armnn/backends/ClWorkloads/ClAdditionFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClAdditionFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClAdditionFloat32Workload.hpp"
+#include "ClAdditionFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
@@ -13,9 +13,9 @@ namespace armnn
{
using namespace armcomputetensorutils;
-void ClAdditionFloat32Workload::Execute() const
+void ClAdditionFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClAdditionFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClAdditionFloatWorkload_Execute");
ClAdditionBaseWorkload::Execute();
}
diff --git a/src/armnn/backends/ClWorkloads/ClAdditionFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClAdditionFloatWorkload.hpp
index 7eac485cfe..03df9cf0a5 100644
--- a/src/armnn/backends/ClWorkloads/ClAdditionFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClAdditionFloatWorkload.hpp
@@ -10,7 +10,7 @@
namespace armnn
{
-class ClAdditionFloat32Workload : public ClAdditionBaseWorkload<DataType::Float16, DataType::Float32>
+class ClAdditionFloatWorkload : public ClAdditionBaseWorkload<DataType::Float16, DataType::Float32>
{
public:
using ClAdditionBaseWorkload<DataType::Float16, DataType::Float32>::ClAdditionBaseWorkload;
diff --git a/src/armnn/backends/ClWorkloads/ClBatchNormalizationFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClBatchNormalizationFloatWorkload.cpp
index 1849c5d411..a5d5b82f38 100644
--- a/src/armnn/backends/ClWorkloads/ClBatchNormalizationFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClBatchNormalizationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClBatchNormalizationFloat32Workload.hpp"
+#include "ClBatchNormalizationFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
@@ -37,7 +37,7 @@ arm_compute::Status ClBatchNormalizationValidate(const TensorInfo& input,
desc.m_Eps);
}
-ClBatchNormalizationFloat32Workload::ClBatchNormalizationFloat32Workload(
+ClBatchNormalizationFloatWorkload::ClBatchNormalizationFloatWorkload(
const BatchNormalizationQueueDescriptor& descriptor, const WorkloadInfo& info)
: FloatWorkload<BatchNormalizationQueueDescriptor>(descriptor, info)
{
@@ -53,7 +53,7 @@ ClBatchNormalizationFloat32Workload::ClBatchNormalizationFloat32Workload(
m_Beta = std::make_unique<arm_compute::CLTensor>();
BuildArmComputeTensor(*m_Beta, m_Data.m_Beta->GetTensorInfo());
- m_Data.ValidateInputsOutputs("ClBatchNormalizationFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("ClBatchNormalizationFloatWorkload", 1, 1);
arm_compute::ICLTensor& input = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ICLTensor& output = static_cast<IClTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -77,13 +77,13 @@ ClBatchNormalizationFloat32Workload::ClBatchNormalizationFloat32Workload(
FreeUnusedTensors();
}
-void ClBatchNormalizationFloat32Workload::Execute() const
+void ClBatchNormalizationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClBatchNormalizationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClBatchNormalizationFloatWorkload_Execute");
m_Layer.run();
}
-void ClBatchNormalizationFloat32Workload::FreeUnusedTensors()
+void ClBatchNormalizationFloatWorkload::FreeUnusedTensors()
{
FreeTensorIfUnused(m_Mean);
FreeTensorIfUnused(m_Variance);
diff --git a/src/armnn/backends/ClWorkloads/ClBatchNormalizationFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClBatchNormalizationFloatWorkload.hpp
index a45614a284..edc0d3c802 100644
--- a/src/armnn/backends/ClWorkloads/ClBatchNormalizationFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClBatchNormalizationFloatWorkload.hpp
@@ -18,10 +18,10 @@ arm_compute::Status ClBatchNormalizationValidate(const TensorInfo& input,
const TensorInfo& gamma,
const BatchNormalizationDescriptor& desc);
-class ClBatchNormalizationFloat32Workload : public FloatWorkload<BatchNormalizationQueueDescriptor>
+class ClBatchNormalizationFloatWorkload : public FloatWorkload<BatchNormalizationQueueDescriptor>
{
public:
- ClBatchNormalizationFloat32Workload(const BatchNormalizationQueueDescriptor& descriptor, const WorkloadInfo& info);
+ ClBatchNormalizationFloatWorkload(const BatchNormalizationQueueDescriptor& descriptor, const WorkloadInfo& info);
using FloatWorkload<BatchNormalizationQueueDescriptor>::FloatWorkload;
void Execute() const override;
diff --git a/src/armnn/backends/ClWorkloads/ClConstantFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClConstantFloatWorkload.cpp
index 58594999a8..842f757ed1 100644
--- a/src/armnn/backends/ClWorkloads/ClConstantFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClConstantFloatWorkload.cpp
@@ -3,13 +3,13 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClConstantFloat32Workload.hpp"
+#include "ClConstantFloatWorkload.hpp"
namespace armnn
{
-void ClConstantFloat32Workload::Execute() const
+void ClConstantFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClConstantFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClConstantFloatWorkload_Execute");
ClBaseConstantWorkload::Execute();
}
diff --git a/src/armnn/backends/ClWorkloads/ClConstantFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClConstantFloatWorkload.hpp
index 11c3fda8db..52b1be46b1 100644
--- a/src/armnn/backends/ClWorkloads/ClConstantFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClConstantFloatWorkload.hpp
@@ -9,7 +9,7 @@
namespace armnn
{
-class ClConstantFloat32Workload : public ClBaseConstantWorkload<DataType::Float16, DataType::Float32>
+class ClConstantFloatWorkload : public ClBaseConstantWorkload<DataType::Float16, DataType::Float32>
{
public:
using ClBaseConstantWorkload<DataType::Float16, DataType::Float32>::ClBaseConstantWorkload;
diff --git a/src/armnn/backends/ClWorkloads/ClConvolution2dFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClConvolution2dFloatWorkload.cpp
index 9ac31df5c1..375ace277a 100644
--- a/src/armnn/backends/ClWorkloads/ClConvolution2dFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClConvolution2dFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClConvolution2dFloat32Workload.hpp"
+#include "ClConvolution2dFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
@@ -13,7 +13,7 @@ namespace armnn
{
using namespace armcomputetensorutils;
-ClConvolution2dFloat32Workload::ClConvolution2dFloat32Workload(const Convolution2dQueueDescriptor& descriptor,
+ClConvolution2dFloatWorkload::ClConvolution2dFloatWorkload(const Convolution2dQueueDescriptor& descriptor,
const WorkloadInfo& info, std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
: FloatWorkload<Convolution2dQueueDescriptor>(descriptor, info)
, m_ConvolutionLayer(memoryManager)
@@ -63,14 +63,14 @@ ClConvolution2dFloat32Workload::ClConvolution2dFloat32Workload(const Convolution
FreeUnusedTensors();
}
-void ClConvolution2dFloat32Workload::Execute() const
+void ClConvolution2dFloatWorkload::Execute() const
{
ARMNN_SCOPED_PROFILING_EVENT_CL("ClConvolution2dFloat32Workload_Execute");
m_ConvolutionLayer.run();
}
-void ClConvolution2dFloat32Workload::FreeUnusedTensors()
+void ClConvolution2dFloatWorkload::FreeUnusedTensors()
{
FreeTensorIfUnused(m_KernelTensor);
FreeTensorIfUnused(m_BiasTensor);
diff --git a/src/armnn/backends/ClWorkloads/ClConvolution2dFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClConvolution2dFloatWorkload.hpp
index 51c21aec32..b0af0951de 100644
--- a/src/armnn/backends/ClWorkloads/ClConvolution2dFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClConvolution2dFloatWorkload.hpp
@@ -14,11 +14,11 @@
namespace armnn
{
-class ClConvolution2dFloat32Workload : public FloatWorkload<Convolution2dQueueDescriptor>
+class ClConvolution2dFloatWorkload : public FloatWorkload<Convolution2dQueueDescriptor>
{
public:
- ClConvolution2dFloat32Workload(const Convolution2dQueueDescriptor& descriptor, const WorkloadInfo& info,
- std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
+ ClConvolution2dFloatWorkload(const Convolution2dQueueDescriptor& descriptor, const WorkloadInfo& info,
+ std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
void Execute() const override;
private:
diff --git a/src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.cpp
index 96d97ad4ea..79aef3ee4e 100644
--- a/src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClDepthwiseConvolutionFloat32Workload.hpp"
+#include "ClDepthwiseConvolutionFloatWorkload.hpp"
#include "backends/ClWorkloadUtils.hpp"
#include "backends/CpuTensorHandle.hpp"
@@ -11,7 +11,7 @@
namespace armnn
{
-ClDepthwiseConvolutionFloat32Workload::ClDepthwiseConvolutionFloat32Workload(
+ClDepthwiseConvolutionFloatWorkload::ClDepthwiseConvolutionFloatWorkload(
const DepthwiseConvolution2dQueueDescriptor& descriptor,
const WorkloadInfo& info)
: ClDepthwiseConvolutionBaseWorkload(descriptor, info)
@@ -27,9 +27,9 @@ ClDepthwiseConvolutionFloat32Workload::ClDepthwiseConvolutionFloat32Workload(
FreeUnusedTensors();
}
-void ClDepthwiseConvolutionFloat32Workload::Execute() const
+void ClDepthwiseConvolutionFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClDepthwiseConvolutionFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClDepthwiseConvolutionFloatWorkload_Execute");
BOOST_ASSERT(m_DepthwiseConvolutionLayer);
m_DepthwiseConvolutionLayer->run();
diff --git a/src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.hpp
index 669fd928b5..2107a7ec61 100644
--- a/src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClDepthwiseConvolutionFloatWorkload.hpp
@@ -12,12 +12,12 @@
namespace armnn
{
-class ClDepthwiseConvolutionFloat32Workload : public ClDepthwiseConvolutionBaseWorkload<DataType::Float16,
- DataType::Float32>
+class ClDepthwiseConvolutionFloatWorkload : public ClDepthwiseConvolutionBaseWorkload<DataType::Float16,
+ DataType::Float32>
{
public:
- ClDepthwiseConvolutionFloat32Workload(const DepthwiseConvolution2dQueueDescriptor& descriptor,
- const WorkloadInfo& info);
+ ClDepthwiseConvolutionFloatWorkload(const DepthwiseConvolution2dQueueDescriptor& descriptor,
+ const WorkloadInfo& info);
void Execute() const override;
};
diff --git a/src/armnn/backends/ClWorkloads/ClFloorFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClFloorFloatWorkload.cpp
index da71c50305..3ec7e753bb 100644
--- a/src/armnn/backends/ClWorkloads/ClFloorFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClFloorFloatWorkload.cpp
@@ -3,16 +3,16 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClFloorFloat32Workload.hpp"
+#include "ClFloorFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
namespace armnn
{
-ClFloorFloat32Workload::ClFloorFloat32Workload(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info)
+ClFloorFloatWorkload::ClFloorFloatWorkload(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info)
: FloatWorkload<FloorQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("ClFloorFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("ClFloorFloatWorkload", 1, 1);
arm_compute::ICLTensor& input = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ICLTensor& output = static_cast<IClTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -20,9 +20,9 @@ ClFloorFloat32Workload::ClFloorFloat32Workload(const FloorQueueDescriptor& descr
m_Layer.configure(&input, &output);
}
-void ClFloorFloat32Workload::Execute() const
+void ClFloorFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClFloorFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClFloorFloatWorkload_Execute");
m_Layer.run();
}
diff --git a/src/armnn/backends/ClWorkloads/ClFloorFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClFloorFloatWorkload.hpp
index bd7f3032fc..965ae06542 100644
--- a/src/armnn/backends/ClWorkloads/ClFloorFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClFloorFloatWorkload.hpp
@@ -10,10 +10,10 @@
namespace armnn
{
-class ClFloorFloat32Workload : public FloatWorkload<FloorQueueDescriptor>
+class ClFloorFloatWorkload : public FloatWorkload<FloorQueueDescriptor>
{
public:
- ClFloorFloat32Workload(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info);
+ ClFloorFloatWorkload(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info);
void Execute() const override;
diff --git a/src/armnn/backends/ClWorkloads/ClFullyConnectedFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClFullyConnectedFloatWorkload.cpp
index 5014dd27ca..9774368597 100644
--- a/src/armnn/backends/ClWorkloads/ClFullyConnectedFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClFullyConnectedFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClFullyConnectedFloat32Workload.hpp"
+#include "ClFullyConnectedFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
@@ -42,7 +42,7 @@ arm_compute::Status ClFullyConnectedWorkloadValidate(const TensorInfo& input,
fullyConnectedLayerInfo);
}
-ClFullyConnectedFloat32Workload::ClFullyConnectedFloat32Workload(const FullyConnectedQueueDescriptor& descriptor,
+ClFullyConnectedFloatWorkload::ClFullyConnectedFloatWorkload(const FullyConnectedQueueDescriptor& descriptor,
const WorkloadInfo& info, std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
: FloatWorkload<FullyConnectedQueueDescriptor>(descriptor, info)
, m_FullyConnectedLayer(memoryManager)
@@ -56,7 +56,7 @@ ClFullyConnectedFloat32Workload::ClFullyConnectedFloat32Workload(const FullyConn
BuildArmComputeTensor(*m_BiasesTensor, m_Data.m_Bias->GetTensorInfo());
}
- m_Data.ValidateInputsOutputs("ClFullyConnectedFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("ClFullyConnectedFloatWorkload", 1, 1);
arm_compute::ICLTensor& input = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ICLTensor& output = static_cast<IClTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -80,13 +80,13 @@ ClFullyConnectedFloat32Workload::ClFullyConnectedFloat32Workload(const FullyConn
FreeUnusedTensors();
}
-void ClFullyConnectedFloat32Workload::Execute() const
+void ClFullyConnectedFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClFullyConnectedFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClFullyConnectedFloatWorkload_Execute");
m_FullyConnectedLayer.run();
}
-void ClFullyConnectedFloat32Workload::FreeUnusedTensors()
+void ClFullyConnectedFloatWorkload::FreeUnusedTensors()
{
FreeTensorIfUnused(m_WeightsTensor);
FreeTensorIfUnused(m_BiasesTensor);
diff --git a/src/armnn/backends/ClWorkloads/ClFullyConnectedFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClFullyConnectedFloatWorkload.hpp
index f580e580c6..1832e4f957 100644
--- a/src/armnn/backends/ClWorkloads/ClFullyConnectedFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClFullyConnectedFloatWorkload.hpp
@@ -20,12 +20,12 @@ arm_compute::Status ClFullyConnectedWorkloadValidate(const TensorInfo& input,
const TensorInfo& biases,
const FullyConnectedDescriptor& descriptor);
-class ClFullyConnectedFloat32Workload : public armnn::FloatWorkload<armnn::FullyConnectedQueueDescriptor>
+class ClFullyConnectedFloatWorkload : public armnn::FloatWorkload<armnn::FullyConnectedQueueDescriptor>
{
public:
- ClFullyConnectedFloat32Workload(const armnn::FullyConnectedQueueDescriptor& descriptor,
- const armnn::WorkloadInfo& info,
- std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
+ ClFullyConnectedFloatWorkload(const armnn::FullyConnectedQueueDescriptor& descriptor,
+ const armnn::WorkloadInfo& info,
+ std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
using armnn::FloatWorkload<armnn::FullyConnectedQueueDescriptor>::m_Data;
void Execute() const override;
diff --git a/src/armnn/backends/ClWorkloads/ClL2NormalizationFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClL2NormalizationFloatWorkload.cpp
index 628e38d3da..c79ed9716c 100644
--- a/src/armnn/backends/ClWorkloads/ClL2NormalizationFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClL2NormalizationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClL2NormalizationFloat32Workload.hpp"
+#include "ClL2NormalizationFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ArmComputeUtils.hpp"
@@ -24,20 +24,20 @@ arm_compute::Status ClL2NormalizationWorkloadValidate(const TensorInfo& input,
return arm_compute::CLNormalizationLayer::validate(&aclInput, &aclOutput, normalizationInfo);
}
-ClL2NormalizationFloat32Workload::ClL2NormalizationFloat32Workload(const L2NormalizationQueueDescriptor& descriptor,
- const WorkloadInfo& info)
+ClL2NormalizationFloatWorkload::ClL2NormalizationFloatWorkload(const L2NormalizationQueueDescriptor& descriptor,
+ const WorkloadInfo& info)
: FloatWorkload<L2NormalizationQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("ClL2NormalizationFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("ClL2NormalizationFloatWorkload", 1, 1);
arm_compute::ICLTensor& input = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ICLTensor& output = static_cast<IClTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
m_Layer.configure(&input, &output, CreateAclNormalizationLayerInfoForL2Normalization(info.m_InputTensorInfos[0]));
}
-void ClL2NormalizationFloat32Workload::Execute() const
+void ClL2NormalizationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClL2NormalizationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClL2NormalizationFloatWorkload_Execute");
m_Layer.run();
}
diff --git a/src/armnn/backends/ClWorkloads/ClL2NormalizationFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClL2NormalizationFloatWorkload.hpp
index bf898e31f7..a247c68a6b 100644
--- a/src/armnn/backends/ClWorkloads/ClL2NormalizationFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClL2NormalizationFloatWorkload.hpp
@@ -13,10 +13,10 @@ namespace armnn
arm_compute::Status ClL2NormalizationWorkloadValidate(const TensorInfo& input,
const TensorInfo& output);
-class ClL2NormalizationFloat32Workload : public FloatWorkload<L2NormalizationQueueDescriptor>
+class ClL2NormalizationFloatWorkload : public FloatWorkload<L2NormalizationQueueDescriptor>
{
public:
- ClL2NormalizationFloat32Workload(const L2NormalizationQueueDescriptor& descriptor, const WorkloadInfo& info);
+ ClL2NormalizationFloatWorkload(const L2NormalizationQueueDescriptor& descriptor, const WorkloadInfo& info);
void Execute() const override;
diff --git a/src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.hpp
deleted file mode 100644
index e2358ad10d..0000000000
--- a/src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.hpp
+++ /dev/null
@@ -1,67 +0,0 @@
-//
-// Copyright © 2017 Arm Ltd. All rights reserved.
-// See LICENSE file in the project root for full license information.
-//
-
-#pragma once
-
-#include "backends/ClWorkloadUtils.hpp"
-#include "backends/Workload.hpp"
-#include "backends/WorkloadData.hpp"
-
-namespace armnn
-{
-
-class ClLstmFloat32Workload : public FloatWorkload<LstmQueueDescriptor>
-{
-public:
- ClLstmFloat32Workload(const LstmQueueDescriptor& descriptor, const WorkloadInfo& info);
- void Execute() const override;
-
-private:
- mutable arm_compute::CLLSTMLayer m_LstmLayer;
-
- std::unique_ptr<arm_compute::CLTensor> m_InputToInputWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_InputToForgetWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_InputToCellWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_InputToOutputWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_RecurrentToInputWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_RecurrentToForgetWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_RecurrentToCellWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_RecurrentToOutputWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_CellToInputWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_CellToForgetWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_CellToOutputWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_InputGateBiasTensor;
- std::unique_ptr<arm_compute::CLTensor> m_ForgetGateBiasTensor;
- std::unique_ptr<arm_compute::CLTensor> m_CellBiasTensor;
- std::unique_ptr<arm_compute::CLTensor> m_OutputGateBiasTensor;
- std::unique_ptr<arm_compute::CLTensor> m_ProjectionWeightsTensor;
- std::unique_ptr<arm_compute::CLTensor> m_ProjectionBiasTensor;
-
- std::unique_ptr<arm_compute::CLTensor> m_ScratchBuffer;
-
- void FreeUnusedTensors();
-};
-
-arm_compute::Status ClLstmFloat32WorkloadValidate(const TensorInfo& input, const TensorInfo& outputStateIn,
- const TensorInfo& cellStateIn, const TensorInfo& scratchBuffer,
- const TensorInfo& outputStateOut, const TensorInfo& cellStateOut,
- const TensorInfo& output, const LstmDescriptor &descriptor,
- const TensorInfo& inputToForgetWeights,
- const TensorInfo& inputToCellWeights,
- const TensorInfo& inputToOutputWeights,
- const TensorInfo& recurrentToForgetWeights,
- const TensorInfo& recurrentToCellWeights,
- const TensorInfo& recurrentToOutputWeights,
- const TensorInfo& forgetGateBias, const TensorInfo& cellBias,
- const TensorInfo& outputGateBias,
- const TensorInfo* inputToInputWeights,
- const TensorInfo* recurrentToInputWeights,
- const TensorInfo* cellToInputWeights,
- const TensorInfo* inputGateBias,
- const TensorInfo* projectionWeights,
- const TensorInfo* projectionBias,
- const TensorInfo* cellToForgetWeights,
- const TensorInfo* cellToOutputWeights);
-} //namespace armnn
diff --git a/src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.cpp
index db5c303854..bfa439bba9 100644
--- a/src/armnn/backends/ClWorkloads/ClLstmFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClLstmFloat32Workload.hpp"
+#include "ClLstmFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
@@ -14,7 +14,7 @@ namespace armnn
{
using namespace armcomputetensorutils;
-ClLstmFloat32Workload::ClLstmFloat32Workload(const LstmQueueDescriptor &descriptor, const WorkloadInfo &info)
+ClLstmFloatWorkload::ClLstmFloatWorkload(const LstmQueueDescriptor &descriptor, const WorkloadInfo &info)
: FloatWorkload<LstmQueueDescriptor>(descriptor, info)
{
arm_compute::LSTMParams<arm_compute::ICLTensor> lstm_param;
@@ -228,31 +228,31 @@ ClLstmFloat32Workload::ClLstmFloat32Workload(const LstmQueueDescriptor &descript
FreeUnusedTensors();
}
-void ClLstmFloat32Workload::Execute() const
+void ClLstmFloatWorkload::Execute() const
{
m_LstmLayer.run();
}
-arm_compute::Status ClLstmFloat32WorkloadValidate(const TensorInfo& input, const TensorInfo& outputStateIn,
- const TensorInfo& cellStateIn, const TensorInfo& scratchBuffer,
- const TensorInfo& outputStateOut, const TensorInfo& cellStateOut,
- const TensorInfo& output, const LstmDescriptor& descriptor,
- const TensorInfo& inputToForgetWeights,
- const TensorInfo& inputToCellWeights,
- const TensorInfo& inputToOutputWeights,
- const TensorInfo& recurrentToForgetWeights,
- const TensorInfo& recurrentToCellWeights,
- const TensorInfo& recurrentToOutputWeights,
- const TensorInfo& forgetGateBias, const TensorInfo& cellBias,
- const TensorInfo& outputGateBias,
- const TensorInfo* inputToInputWeights,
- const TensorInfo* recurrentToInputWeights,
- const TensorInfo* cellToInputWeights,
- const TensorInfo* inputGateBias,
- const TensorInfo* projectionWeights,
- const TensorInfo* projectionBias,
- const TensorInfo* cellToForgetWeights,
- const TensorInfo* cellToOutputWeights)
+arm_compute::Status ClLstmFloatWorkloadValidate(const TensorInfo& input, const TensorInfo& outputStateIn,
+ const TensorInfo& cellStateIn, const TensorInfo& scratchBuffer,
+ const TensorInfo& outputStateOut, const TensorInfo& cellStateOut,
+ const TensorInfo& output, const LstmDescriptor& descriptor,
+ const TensorInfo& inputToForgetWeights,
+ const TensorInfo& inputToCellWeights,
+ const TensorInfo& inputToOutputWeights,
+ const TensorInfo& recurrentToForgetWeights,
+ const TensorInfo& recurrentToCellWeights,
+ const TensorInfo& recurrentToOutputWeights,
+ const TensorInfo& forgetGateBias, const TensorInfo& cellBias,
+ const TensorInfo& outputGateBias,
+ const TensorInfo* inputToInputWeights,
+ const TensorInfo* recurrentToInputWeights,
+ const TensorInfo* cellToInputWeights,
+ const TensorInfo* inputGateBias,
+ const TensorInfo* projectionWeights,
+ const TensorInfo* projectionBias,
+ const TensorInfo* cellToForgetWeights,
+ const TensorInfo* cellToOutputWeights)
{
arm_compute::LSTMParams<arm_compute::ITensorInfo> lstm_params_info;
@@ -380,7 +380,7 @@ arm_compute::Status ClLstmFloat32WorkloadValidate(const TensorInfo& input, const
cell_threshold, projection_threshold);
}
-void ClLstmFloat32Workload::FreeUnusedTensors()
+void ClLstmFloatWorkload::FreeUnusedTensors()
{
FreeTensorIfUnused(m_InputToInputWeightsTensor);
FreeTensorIfUnused(m_InputToForgetWeightsTensor);
diff --git a/src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.hpp b/src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.hpp
new file mode 100644
index 0000000000..47473b2fdf
--- /dev/null
+++ b/src/armnn/backends/ClWorkloads/ClLstmFloatWorkload.hpp
@@ -0,0 +1,67 @@
+//
+// Copyright © 2017 Arm Ltd. All rights reserved.
+// See LICENSE file in the project root for full license information.
+//
+
+#pragma once
+
+#include "backends/ClWorkloadUtils.hpp"
+#include "backends/Workload.hpp"
+#include "backends/WorkloadData.hpp"
+
+namespace armnn
+{
+
+class ClLstmFloatWorkload : public FloatWorkload<LstmQueueDescriptor>
+{
+public:
+ ClLstmFloatWorkload(const LstmQueueDescriptor& descriptor, const WorkloadInfo& info);
+ void Execute() const override;
+
+private:
+ mutable arm_compute::CLLSTMLayer m_LstmLayer;
+
+ std::unique_ptr<arm_compute::CLTensor> m_InputToInputWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_InputToForgetWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_InputToCellWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_InputToOutputWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_RecurrentToInputWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_RecurrentToForgetWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_RecurrentToCellWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_RecurrentToOutputWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_CellToInputWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_CellToForgetWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_CellToOutputWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_InputGateBiasTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_ForgetGateBiasTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_CellBiasTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_OutputGateBiasTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_ProjectionWeightsTensor;
+ std::unique_ptr<arm_compute::CLTensor> m_ProjectionBiasTensor;
+
+ std::unique_ptr<arm_compute::CLTensor> m_ScratchBuffer;
+
+ void FreeUnusedTensors();
+};
+
+arm_compute::Status ClLstmFloatWorkloadValidate(const TensorInfo& input, const TensorInfo& outputStateIn,
+ const TensorInfo& cellStateIn, const TensorInfo& scratchBuffer,
+ const TensorInfo& outputStateOut, const TensorInfo& cellStateOut,
+ const TensorInfo& output, const LstmDescriptor &descriptor,
+ const TensorInfo& inputToForgetWeights,
+ const TensorInfo& inputToCellWeights,
+ const TensorInfo& inputToOutputWeights,
+ const TensorInfo& recurrentToForgetWeights,
+ const TensorInfo& recurrentToCellWeights,
+ const TensorInfo& recurrentToOutputWeights,
+ const TensorInfo& forgetGateBias, const TensorInfo& cellBias,
+ const TensorInfo& outputGateBias,
+ const TensorInfo* inputToInputWeights,
+ const TensorInfo* recurrentToInputWeights,
+ const TensorInfo* cellToInputWeights,
+ const TensorInfo* inputGateBias,
+ const TensorInfo* projectionWeights,
+ const TensorInfo* projectionBias,
+ const TensorInfo* cellToForgetWeights,
+ const TensorInfo* cellToOutputWeights);
+} //namespace armnn
diff --git a/src/armnn/backends/ClWorkloads/ClMergerFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClMergerFloatWorkload.cpp
index 89e7690a36..3d8a183b70 100644
--- a/src/armnn/backends/ClWorkloads/ClMergerFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClMergerFloatWorkload.cpp
@@ -3,15 +3,15 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClMergerFloat32Workload.hpp"
+#include "ClMergerFloatWorkload.hpp"
namespace armnn
{
-void ClMergerFloat32Workload::Execute() const
+void ClMergerFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClMergerFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClMergerFloatWorkload_Execute");
ClBaseMergerWorkload::Execute();
}
diff --git a/src/armnn/backends/ClWorkloads/ClMergerFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClMergerFloatWorkload.hpp
index 3cafa23c1e..e89f96eb86 100644
--- a/src/armnn/backends/ClWorkloads/ClMergerFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClMergerFloatWorkload.hpp
@@ -10,7 +10,7 @@
namespace armnn
{
-class ClMergerFloat32Workload : public ClBaseMergerWorkload<DataType::Float16, DataType::Float32>
+class ClMergerFloatWorkload : public ClBaseMergerWorkload<DataType::Float16, DataType::Float32>
{
public:
using ClBaseMergerWorkload<DataType::Float16, DataType::Float32>::ClBaseMergerWorkload;
diff --git a/src/armnn/backends/ClWorkloads/ClMultiplicationFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClMultiplicationFloatWorkload.cpp
index 7aa33146f3..e161a0a8fe 100644
--- a/src/armnn/backends/ClWorkloads/ClMultiplicationFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClMultiplicationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClMultiplicationFloat32Workload.hpp"
+#include "ClMultiplicationFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
@@ -30,11 +30,11 @@ arm_compute::Status ClMultiplicationWorkloadValidate(const TensorInfo& input0,
}
-ClMultiplicationFloat32Workload::ClMultiplicationFloat32Workload(const MultiplicationQueueDescriptor& descriptor,
+ClMultiplicationFloatWorkload::ClMultiplicationFloatWorkload(const MultiplicationQueueDescriptor& descriptor,
const WorkloadInfo& info)
: FloatWorkload<MultiplicationQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("ClMultiplicationFloat32Workload", 2, 1);
+ m_Data.ValidateInputsOutputs("ClMultiplicationFloatWorkload", 2, 1);
arm_compute::ICLTensor& input0 = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ICLTensor& input1 = static_cast<IClTensorHandle*>(m_Data.m_Inputs[1])->GetTensor();
@@ -48,9 +48,9 @@ ClMultiplicationFloat32Workload::ClMultiplicationFloat32Workload(const Multiplic
arm_compute::RoundingPolicy::TO_NEAREST_EVEN);
}
-void ClMultiplicationFloat32Workload::Execute() const
+void ClMultiplicationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClMultiplicationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClMultiplicationFloatWorkload_Execute");
// Executes the layer.
m_PixelWiseMultiplication.run();
diff --git a/src/armnn/backends/ClWorkloads/ClMultiplicationFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClMultiplicationFloatWorkload.hpp
index 0d6199047d..0da8451378 100644
--- a/src/armnn/backends/ClWorkloads/ClMultiplicationFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClMultiplicationFloatWorkload.hpp
@@ -14,10 +14,10 @@ arm_compute::Status ClMultiplicationWorkloadValidate(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output);
-class ClMultiplicationFloat32Workload : public FloatWorkload<MultiplicationQueueDescriptor>
+class ClMultiplicationFloatWorkload : public FloatWorkload<MultiplicationQueueDescriptor>
{
public:
- ClMultiplicationFloat32Workload(const MultiplicationQueueDescriptor& descriptor, const WorkloadInfo& info);
+ ClMultiplicationFloatWorkload(const MultiplicationQueueDescriptor& descriptor, const WorkloadInfo& info);
using FloatWorkload<MultiplicationQueueDescriptor>::FloatWorkload;
void Execute() const override;
diff --git a/src/armnn/backends/ClWorkloads/ClNormalizationFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClNormalizationFloatWorkload.cpp
index d23d6e11bd..bd486e6330 100644
--- a/src/armnn/backends/ClWorkloads/ClNormalizationFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClNormalizationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClNormalizationFloat32Workload.hpp"
+#include "ClNormalizationFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ClLayerSupport.hpp"
@@ -25,11 +25,11 @@ arm_compute::Status ClNormalizationWorkloadValidate(const TensorInfo& input, con
return arm_compute::CLNormalizationLayer::validate(&aclInputInfo, &aclOutputInfo, layerInfo);
}
-ClNormalizationFloat32Workload::ClNormalizationFloat32Workload(const NormalizationQueueDescriptor& descriptor,
- const WorkloadInfo& info)
+ClNormalizationFloatWorkload::ClNormalizationFloatWorkload(const NormalizationQueueDescriptor& descriptor,
+ const WorkloadInfo& info)
: FloatWorkload<NormalizationQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("ClNormalizationFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("ClNormalizationFloatWorkload", 1, 1);
arm_compute::ICLTensor& input = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ICLTensor& output = static_cast<IClTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -40,9 +40,9 @@ ClNormalizationFloat32Workload::ClNormalizationFloat32Workload(const Normalizati
m_NormalizationLayer.configure(&input, &output, normalizationInfo);
};
-void ClNormalizationFloat32Workload::Execute() const
+void ClNormalizationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClNormalizationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClNormalizationFloatWorkload_Execute");
m_NormalizationLayer.run();
}
diff --git a/src/armnn/backends/ClWorkloads/ClNormalizationFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClNormalizationFloatWorkload.hpp
index e8ab0b9a18..4f49e93eab 100644
--- a/src/armnn/backends/ClWorkloads/ClNormalizationFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClNormalizationFloatWorkload.hpp
@@ -14,15 +14,14 @@ arm_compute::Status ClNormalizationWorkloadValidate(const TensorInfo& input,
const TensorInfo& output,
const NormalizationDescriptor& descriptor);
-class ClNormalizationFloat32Workload : public FloatWorkload<NormalizationQueueDescriptor>
+class ClNormalizationFloatWorkload : public FloatWorkload<NormalizationQueueDescriptor>
{
public:
- ClNormalizationFloat32Workload(const NormalizationQueueDescriptor& descriptor, const WorkloadInfo& info);
+ ClNormalizationFloatWorkload(const NormalizationQueueDescriptor& descriptor, const WorkloadInfo& info);
void Execute() const override;
private:
mutable arm_compute::CLNormalizationLayer m_NormalizationLayer;
};
-} //namespace armnn
-
+} //namespace armnn \ No newline at end of file
diff --git a/src/armnn/backends/ClWorkloads/ClPooling2dFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClPooling2dFloatWorkload.cpp
index 3a5b8ca526..9979e2fd7b 100644
--- a/src/armnn/backends/ClWorkloads/ClPooling2dFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClPooling2dFloatWorkload.cpp
@@ -3,20 +3,20 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClPooling2dFloat32Workload.hpp"
+#include "ClPooling2dFloatWorkload.hpp"
namespace armnn
{
-ClPooling2dFloat32Workload::ClPooling2dFloat32Workload(const Pooling2dQueueDescriptor& descriptor,
+ClPooling2dFloatWorkload::ClPooling2dFloatWorkload(const Pooling2dQueueDescriptor& descriptor,
const WorkloadInfo& info)
- : ClPooling2dBaseWorkload<DataType::Float16, DataType::Float32>(descriptor, info, "ClPooling2dFloat32Workload")
+ : ClPooling2dBaseWorkload<DataType::Float16, DataType::Float32>(descriptor, info, "ClPooling2dFloatWorkload")
{
}
-void ClPooling2dFloat32Workload::Execute() const
+void ClPooling2dFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClPooling2dFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClPooling2dFloatWorkload_Execute");
m_PoolingLayer.run();
}
diff --git a/src/armnn/backends/ClWorkloads/ClPooling2dFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClPooling2dFloatWorkload.hpp
index ad189bdb52..0b73abb70c 100644
--- a/src/armnn/backends/ClWorkloads/ClPooling2dFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClPooling2dFloatWorkload.hpp
@@ -10,10 +10,10 @@
namespace armnn
{
-class ClPooling2dFloat32Workload : public ClPooling2dBaseWorkload<DataType::Float16, DataType::Float32>
+class ClPooling2dFloatWorkload : public ClPooling2dBaseWorkload<DataType::Float16, DataType::Float32>
{
public:
- ClPooling2dFloat32Workload(const Pooling2dQueueDescriptor& descriptor, const WorkloadInfo& info);
+ ClPooling2dFloatWorkload(const Pooling2dQueueDescriptor& descriptor, const WorkloadInfo& info);
void Execute() const override;
};
diff --git a/src/armnn/backends/ClWorkloads/ClReshapeFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClReshapeFloatWorkload.cpp
index 05fba222ac..645544b75d 100644
--- a/src/armnn/backends/ClWorkloads/ClReshapeFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClReshapeFloatWorkload.cpp
@@ -3,17 +3,17 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClReshapeFloat32Workload.hpp"
+#include "ClReshapeFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
namespace armnn
{
-ClReshapeFloat32Workload::ClReshapeFloat32Workload(const ReshapeQueueDescriptor& descriptor, const WorkloadInfo& info)
+ClReshapeFloatWorkload::ClReshapeFloatWorkload(const ReshapeQueueDescriptor& descriptor, const WorkloadInfo& info)
: FloatWorkload<ReshapeQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("ClReshapeFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("ClReshapeFloatWorkload", 1, 1);
arm_compute::ICLTensor& input = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ICLTensor& output = static_cast<IClTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -21,9 +21,9 @@ ClReshapeFloat32Workload::ClReshapeFloat32Workload(const ReshapeQueueDescriptor&
m_Layer.configure(&input, &output);
}
-void ClReshapeFloat32Workload::Execute() const
+void ClReshapeFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClReshapeFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClReshapeFloatWorkload_Execute");
m_Layer.run();
}
diff --git a/src/armnn/backends/ClWorkloads/ClReshapeFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClReshapeFloatWorkload.hpp
index 0eb4d08da0..18392af99b 100644
--- a/src/armnn/backends/ClWorkloads/ClReshapeFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClReshapeFloatWorkload.hpp
@@ -10,10 +10,10 @@
namespace armnn
{
-class ClReshapeFloat32Workload : public FloatWorkload<ReshapeQueueDescriptor>
+class ClReshapeFloatWorkload : public FloatWorkload<ReshapeQueueDescriptor>
{
public:
- ClReshapeFloat32Workload(const ReshapeQueueDescriptor& descriptor, const WorkloadInfo& info);
+ ClReshapeFloatWorkload(const ReshapeQueueDescriptor& descriptor, const WorkloadInfo& info);
void Execute() const override;
diff --git a/src/armnn/backends/ClWorkloads/ClResizeBilinearFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClResizeBilinearFloatWorkload.cpp
index abef682611..e7e2f70e40 100644
--- a/src/armnn/backends/ClWorkloads/ClResizeBilinearFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClResizeBilinearFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClResizeBilinearFloat32Workload.hpp"
+#include "ClResizeBilinearFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ClLayerSupport.hpp"
@@ -12,11 +12,11 @@
namespace armnn
{
-ClResizeBilinearFloat32Workload::ClResizeBilinearFloat32Workload(const ResizeBilinearQueueDescriptor& descriptor,
+ClResizeBilinearFloatWorkload::ClResizeBilinearFloatWorkload(const ResizeBilinearQueueDescriptor& descriptor,
const WorkloadInfo& info)
: FloatWorkload<ResizeBilinearQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("ClResizeBilinearFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("ClResizeBilinearFloatWorkload", 1, 1);
arm_compute::ICLTensor& input = static_cast<IClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ICLTensor& output = static_cast<IClTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -26,9 +26,9 @@ ClResizeBilinearFloat32Workload::ClResizeBilinearFloat32Workload(const ResizeBil
arm_compute::SamplingPolicy::TOP_LEFT);
};
-void ClResizeBilinearFloat32Workload::Execute() const
+void ClResizeBilinearFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClResizeBilinearFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClResizeBilinearFloatWorkload_Execute");
m_ResizeBilinearLayer.run();
}
diff --git a/src/armnn/backends/ClWorkloads/ClResizeBilinearFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClResizeBilinearFloatWorkload.hpp
index 81c0566bb3..7527cc68db 100644
--- a/src/armnn/backends/ClWorkloads/ClResizeBilinearFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClResizeBilinearFloatWorkload.hpp
@@ -10,10 +10,10 @@
namespace armnn
{
-class ClResizeBilinearFloat32Workload : public FloatWorkload<ResizeBilinearQueueDescriptor>
+class ClResizeBilinearFloatWorkload : public FloatWorkload<ResizeBilinearQueueDescriptor>
{
public:
- ClResizeBilinearFloat32Workload(const ResizeBilinearQueueDescriptor& descriptor, const WorkloadInfo& info);
+ ClResizeBilinearFloatWorkload(const ResizeBilinearQueueDescriptor& descriptor, const WorkloadInfo& info);
void Execute() const override;
private:
diff --git a/src/armnn/backends/ClWorkloads/ClSoftmaxFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClSoftmaxFloatWorkload.cpp
index 08247bc593..2816d80f74 100644
--- a/src/armnn/backends/ClWorkloads/ClSoftmaxFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClSoftmaxFloatWorkload.cpp
@@ -3,28 +3,28 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClSoftmaxFloat32Workload.hpp"
+#include "ClSoftmaxFloatWorkload.hpp"
#include "backends/ClTensorHandle.hpp"
#include "backends/CpuTensorHandle.hpp"
namespace armnn
{
-ClSoftmaxFloat32Workload::ClSoftmaxFloat32Workload(const SoftmaxQueueDescriptor& descriptor, const WorkloadInfo& info,
+ClSoftmaxFloatWorkload::ClSoftmaxFloatWorkload(const SoftmaxQueueDescriptor& descriptor, const WorkloadInfo& info,
std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
: FloatWorkload<SoftmaxQueueDescriptor>(descriptor, info)
, m_SoftmaxLayer(memoryManager)
{
- m_Data.ValidateInputsOutputs("ClSoftmaxFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("ClSoftmaxFloatWorkload", 1, 1);
arm_compute::ICLTensor& input = static_cast<ClTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ICLTensor& output = static_cast<ClTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
m_SoftmaxLayer.configure(&input, &output, m_Data.m_Parameters.m_Beta);
}
-void ClSoftmaxFloat32Workload::Execute() const
+void ClSoftmaxFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClSoftmaxFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClSoftmaxFloatWorkload_Execute");
m_SoftmaxLayer.run();
}
diff --git a/src/armnn/backends/ClWorkloads/ClSoftmaxFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClSoftmaxFloatWorkload.hpp
index 6cad59800b..d50ba6f06b 100644
--- a/src/armnn/backends/ClWorkloads/ClSoftmaxFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClSoftmaxFloatWorkload.hpp
@@ -14,10 +14,10 @@
namespace armnn
{
-class ClSoftmaxFloat32Workload : public FloatWorkload<SoftmaxQueueDescriptor>
+class ClSoftmaxFloatWorkload : public FloatWorkload<SoftmaxQueueDescriptor>
{
public:
- ClSoftmaxFloat32Workload(const SoftmaxQueueDescriptor& descriptor, const WorkloadInfo& info,
+ ClSoftmaxFloatWorkload(const SoftmaxQueueDescriptor& descriptor, const WorkloadInfo& info,
std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
void Execute() const override;
diff --git a/src/armnn/backends/ClWorkloads/ClSplitterFloat32Workload.cpp b/src/armnn/backends/ClWorkloads/ClSplitterFloatWorkload.cpp
index 8a622c6caf..816cc6a1c4 100644
--- a/src/armnn/backends/ClWorkloads/ClSplitterFloat32Workload.cpp
+++ b/src/armnn/backends/ClWorkloads/ClSplitterFloatWorkload.cpp
@@ -3,14 +3,14 @@
// See LICENSE file in the project root for full license information.
//
-#include "ClSplitterFloat32Workload.hpp"
+#include "ClSplitterFloatWorkload.hpp"
namespace armnn
{
-void ClSplitterFloat32Workload::Execute() const
+void ClSplitterFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_CL("ClSplitterFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_CL("ClSplitterFloatWorkload_Execute");
ClBaseSplitterWorkload::Execute();
}
diff --git a/src/armnn/backends/ClWorkloads/ClSplitterFloat32Workload.hpp b/src/armnn/backends/ClWorkloads/ClSplitterFloatWorkload.hpp
index affa9f840f..f795473548 100644
--- a/src/armnn/backends/ClWorkloads/ClSplitterFloat32Workload.hpp
+++ b/src/armnn/backends/ClWorkloads/ClSplitterFloatWorkload.hpp
@@ -10,7 +10,7 @@
namespace armnn
{
-class ClSplitterFloat32Workload : public ClBaseSplitterWorkload<DataType::Float16, DataType::Float32>
+class ClSplitterFloatWorkload : public ClBaseSplitterWorkload<DataType::Float16, DataType::Float32>
{
public:
using ClBaseSplitterWorkload<DataType::Float16, DataType::Float32>::ClBaseSplitterWorkload;
diff --git a/src/armnn/backends/NeonLayerSupport.cpp b/src/armnn/backends/NeonLayerSupport.cpp
index 48b3ccbfa0..fd2372f6f3 100644
--- a/src/armnn/backends/NeonLayerSupport.cpp
+++ b/src/armnn/backends/NeonLayerSupport.cpp
@@ -15,15 +15,15 @@
#include <boost/core/ignore_unused.hpp>
#ifdef ARMCOMPUTENEON_ENABLED
-#include "NeonWorkloads/NeonAdditionFloat32Workload.hpp"
-#include "NeonWorkloads/NeonActivationFloat32Workload.hpp"
-#include "NeonWorkloads/NeonBatchNormalizationFloat32Workload.hpp"
+#include "NeonWorkloads/NeonAdditionFloatWorkload.hpp"
+#include "NeonWorkloads/NeonActivationFloatWorkload.hpp"
+#include "NeonWorkloads/NeonBatchNormalizationFloatWorkload.hpp"
#include "NeonWorkloads/NeonConvolution2dBaseWorkload.hpp"
#include "NeonWorkloads/NeonDepthwiseConvolutionBaseWorkload.hpp"
-#include "NeonWorkloads/NeonL2NormalizationFloat32Workload.hpp"
-#include "NeonWorkloads/NeonMultiplicationFloat32Workload.hpp"
-#include "NeonWorkloads/NeonNormalizationFloat32Workload.hpp"
-#include "NeonWorkloads/NeonFullyConnectedFloat32Workload.hpp"
+#include "NeonWorkloads/NeonL2NormalizationFloatWorkload.hpp"
+#include "NeonWorkloads/NeonMultiplicationFloatWorkload.hpp"
+#include "NeonWorkloads/NeonNormalizationFloatWorkload.hpp"
+#include "NeonWorkloads/NeonFullyConnectedFloatWorkload.hpp"
#include "NeonWorkloads/NeonPermuteWorkload.hpp"
#include "NeonWorkloads/NeonPooling2dBaseWorkload.hpp"
#include "NeonWorkloads/NeonSoftmaxBaseWorkload.hpp"
diff --git a/src/armnn/backends/NeonWorkloadFactory.cpp b/src/armnn/backends/NeonWorkloadFactory.cpp
index 2332b8b845..6e99701037 100644
--- a/src/armnn/backends/NeonWorkloadFactory.cpp
+++ b/src/armnn/backends/NeonWorkloadFactory.cpp
@@ -82,32 +82,32 @@ std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateOutput(const OutputQueueDe
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateActivation(const ActivationQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonActivationFloat32Workload, NeonActivationUint8Workload>(descriptor, info);
+ return MakeWorkload<NeonActivationFloatWorkload, NeonActivationUint8Workload>(descriptor, info);
}
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateSoftmax(const SoftmaxQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonSoftmaxFloat32Workload, NeonSoftmaxUint8Workload>(descriptor, info,
+ return MakeWorkload<NeonSoftmaxFloatWorkload, NeonSoftmaxUint8Workload>(descriptor, info,
m_MemoryManager.GetIntraLayerManager());
}
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateSplitter(const SplitterQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonSplitterFloat32Workload, NeonSplitterUint8Workload>(descriptor, info);
+ return MakeWorkload<NeonSplitterFloatWorkload, NeonSplitterUint8Workload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateMerger(const MergerQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonMergerFloat32Workload, NeonMergerUint8Workload>(descriptor, info);
+ return MakeWorkload<NeonMergerFloatWorkload, NeonMergerUint8Workload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateFullyConnected(
const FullyConnectedQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<NeonFullyConnectedFloat32Workload, NullWorkload>(descriptor, info,
+ return MakeWorkload<NeonFullyConnectedFloatWorkload, NullWorkload>(descriptor, info,
m_MemoryManager.GetIntraLayerManager());
}
@@ -120,40 +120,40 @@ std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreatePermute(const Permu
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreatePooling2d(const Pooling2dQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonPooling2dFloat32Workload, NeonPooling2dUint8Workload>(descriptor, info);
+ return MakeWorkload<NeonPooling2dFloatWorkload, NeonPooling2dUint8Workload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateConvolution2d(
const Convolution2dQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<NeonConvolution2dFloat32Workload, NeonConvolution2dUint8Workload>(descriptor, info,
+ return MakeWorkload<NeonConvolution2dFloatWorkload, NeonConvolution2dUint8Workload>(descriptor, info,
m_MemoryManager.GetIntraLayerManager());
}
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateDepthwiseConvolution2d(
const DepthwiseConvolution2dQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<NeonDepthwiseConvolutionFloat32Workload, NeonDepthwiseConvolutionUint8Workload>(
+ return MakeWorkload<NeonDepthwiseConvolutionFloatWorkload, NeonDepthwiseConvolutionUint8Workload>(
descriptor, info);
}
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateNormalization(
const NormalizationQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<NeonNormalizationFloat32Workload, NullWorkload>(descriptor, info,
+ return MakeWorkload<NeonNormalizationFloatWorkload, NullWorkload>(descriptor, info,
m_MemoryManager.GetIntraLayerManager());
}
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateAddition(const AdditionQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonAdditionFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<NeonAdditionFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateMultiplication(
const MultiplicationQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<NeonMultiplicationFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<NeonMultiplicationFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateDivision(
@@ -165,7 +165,7 @@ std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateDivision(
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateBatchNormalization(
const BatchNormalizationQueueDescriptor& descriptor, const WorkloadInfo& info) const
{
- return MakeWorkload<NeonBatchNormalizationFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<NeonBatchNormalizationFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateMemCopy(const MemCopyQueueDescriptor& descriptor,
@@ -196,32 +196,32 @@ std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateFakeQuantization(
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateL2Normalization(const L2NormalizationQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonL2NormalizationFloat32Workload, NullWorkload>(descriptor, info,
+ return MakeWorkload<NeonL2NormalizationFloatWorkload, NullWorkload>(descriptor, info,
m_MemoryManager.GetIntraLayerManager());
}
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateConstant(const ConstantQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonConstantFloat32Workload, NeonConstantUint8Workload>(descriptor, info);
+ return MakeWorkload<NeonConstantFloatWorkload, NeonConstantUint8Workload>(descriptor, info);
}
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateReshape(const ReshapeQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonReshapeFloat32Workload, NeonReshapeUint8Workload>(descriptor, info);
+ return MakeWorkload<NeonReshapeFloatWorkload, NeonReshapeUint8Workload>(descriptor, info);
}
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateFloor(const FloorQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonFloorFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<NeonFloorFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateLstm(const LstmQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- return MakeWorkload<NeonLstmFloat32Workload, NullWorkload>(descriptor, info);
+ return MakeWorkload<NeonLstmFloatWorkload, NullWorkload>(descriptor, info);
}
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateConvertFp16ToFp32(
diff --git a/src/armnn/backends/NeonWorkloads.hpp b/src/armnn/backends/NeonWorkloads.hpp
index 9619b4e5c9..1948232b53 100644
--- a/src/armnn/backends/NeonWorkloads.hpp
+++ b/src/armnn/backends/NeonWorkloads.hpp
@@ -4,37 +4,37 @@
//
#pragma once
-#include "backends/NeonWorkloads/NeonActivationFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonActivationFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonActivationUint8Workload.hpp"
-#include "backends/NeonWorkloads/NeonAdditionFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonAdditionFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonBaseConstantWorkload.hpp"
#include "backends/NeonWorkloads/NeonBaseMergerWorkload.hpp"
#include "backends/NeonWorkloads/NeonBaseSplitterWorkload.hpp"
-#include "backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.hpp"
-#include "backends/NeonWorkloads/NeonConstantFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.hpp"
+#include "backends/NeonWorkloads/NeonConstantFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonConstantUint8Workload.hpp"
#include "backends/NeonWorkloads/NeonConvertFp16ToFp32Workload.hpp"
#include "backends/NeonWorkloads/NeonConvertFp32ToFp16Workload.hpp"
#include "backends/NeonWorkloads/NeonConvolution2dBaseWorkload.hpp"
-#include "backends/NeonWorkloads/NeonConvolution2dFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonConvolution2dFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonConvolution2dUint8Workload.hpp"
-#include "backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonDepthwiseConvolutionUint8Workload.hpp"
-#include "backends/NeonWorkloads/NeonFloorFloat32Workload.hpp"
-#include "backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.hpp"
-#include "backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.hpp"
-#include "backends/NeonWorkloads/NeonLstmFloat32Workload.hpp"
-#include "backends/NeonWorkloads/NeonMergerFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonFloorFloatWorkload.hpp"
+#include "backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.hpp"
+#include "backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.hpp"
+#include "backends/NeonWorkloads/NeonLstmFloatWorkload.hpp"
+#include "backends/NeonWorkloads/NeonMergerFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonMergerUint8Workload.hpp"
-#include "backends/NeonWorkloads/NeonMultiplicationFloat32Workload.hpp"
-#include "backends/NeonWorkloads/NeonNormalizationFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonMultiplicationFloatWorkload.hpp"
+#include "backends/NeonWorkloads/NeonNormalizationFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonPermuteWorkload.hpp"
#include "backends/NeonWorkloads/NeonPooling2dBaseWorkload.hpp"
-#include "backends/NeonWorkloads/NeonPooling2dFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonPooling2dFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonPooling2dUint8Workload.hpp"
-#include "backends/NeonWorkloads/NeonReshapeFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonReshapeFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonReshapeUint8Workload.hpp"
-#include "backends/NeonWorkloads/NeonSoftmaxFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonSoftmaxFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonSoftmaxUint8Workload.hpp"
-#include "backends/NeonWorkloads/NeonSplitterFloat32Workload.hpp"
+#include "backends/NeonWorkloads/NeonSplitterFloatWorkload.hpp"
#include "backends/NeonWorkloads/NeonSplitterUint8Workload.hpp"
diff --git a/src/armnn/backends/NeonWorkloads/NeonActivationFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonActivationFloatWorkload.cpp
index 711bfceeaf..e3524dced6 100644
--- a/src/armnn/backends/NeonWorkloads/NeonActivationFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonActivationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonActivationFloat32Workload.hpp"
+#include "NeonActivationFloatWorkload.hpp"
#include "backends/ArmComputeUtils.hpp"
@@ -32,11 +32,11 @@ arm_compute::Status NeonActivationWorkloadValidate(const TensorInfo& input,
activationLayerInfo);
}
-NeonActivationFloat32Workload::NeonActivationFloat32Workload(const ActivationQueueDescriptor& descriptor,
- const WorkloadInfo& info)
+NeonActivationFloatWorkload::NeonActivationFloatWorkload(const ActivationQueueDescriptor& descriptor,
+ const WorkloadInfo& info)
: FloatWorkload<ActivationQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("NeonActivationFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonActivationFloatWorkload", 1, 1);
const arm_compute::ActivationLayerInfo activationLayerInfo =
ConvertActivationDescriptorToAclActivationLayerInfo(m_Data.m_Parameters);
@@ -47,9 +47,9 @@ NeonActivationFloat32Workload::NeonActivationFloat32Workload(const ActivationQue
m_ActivationLayer.configure(&input, &output, activationLayerInfo);
}
-void NeonActivationFloat32Workload::Execute() const
+void NeonActivationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonActivationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonActivationFloatWorkload_Execute");
m_ActivationLayer.run();
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonActivationFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonActivationFloatWorkload.hpp
index 0d26b3b39f..c9f4bab5f0 100644
--- a/src/armnn/backends/NeonWorkloads/NeonActivationFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonActivationFloatWorkload.hpp
@@ -14,10 +14,10 @@ arm_compute::Status NeonActivationWorkloadValidate(const TensorInfo& input,
const TensorInfo& output,
const ActivationDescriptor& descriptor);
-class NeonActivationFloat32Workload : public FloatWorkload<ActivationQueueDescriptor>
+class NeonActivationFloatWorkload : public FloatWorkload<ActivationQueueDescriptor>
{
public:
- NeonActivationFloat32Workload(const ActivationQueueDescriptor& descriptor, const WorkloadInfo& info);
+ NeonActivationFloatWorkload(const ActivationQueueDescriptor& descriptor, const WorkloadInfo& info);
void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonAdditionFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonAdditionFloatWorkload.cpp
index f26e42aff9..1c1e65c206 100644
--- a/src/armnn/backends/NeonWorkloads/NeonAdditionFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonAdditionFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonAdditionFloat32Workload.hpp"
+#include "NeonAdditionFloatWorkload.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
#include "backends/CpuTensorHandle.hpp"
@@ -25,11 +25,11 @@ arm_compute::Status NeonAdditionWorkloadValidate(const TensorInfo& input0,
}
-NeonAdditionFloat32Workload::NeonAdditionFloat32Workload(const AdditionQueueDescriptor& descriptor,
- const WorkloadInfo& info)
+NeonAdditionFloatWorkload::NeonAdditionFloatWorkload(const AdditionQueueDescriptor& descriptor,
+ const WorkloadInfo& info)
: FloatWorkload<AdditionQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("NeonAdditionFloat32Workload", 2, 1);
+ m_Data.ValidateInputsOutputs("NeonAdditionFloatWorkload", 2, 1);
arm_compute::ITensor& input1 = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ITensor& input2 = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[1])->GetTensor();
@@ -38,9 +38,9 @@ NeonAdditionFloat32Workload::NeonAdditionFloat32Workload(const AdditionQueueDesc
m_AddLayer.configure(&input1, &input2, &output, arm_compute::ConvertPolicy::SATURATE);
}
-void NeonAdditionFloat32Workload::Execute() const
+void NeonAdditionFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonAdditionFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonAdditionFloatWorkload_Execute");
m_AddLayer.run();
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonAdditionFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonAdditionFloatWorkload.hpp
index dae66bb69d..efab230925 100644
--- a/src/armnn/backends/NeonWorkloads/NeonAdditionFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonAdditionFloatWorkload.hpp
@@ -14,10 +14,10 @@ arm_compute::Status NeonAdditionWorkloadValidate(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output);
-class NeonAdditionFloat32Workload : public FloatWorkload<AdditionQueueDescriptor>
+class NeonAdditionFloatWorkload : public FloatWorkload<AdditionQueueDescriptor>
{
public:
- NeonAdditionFloat32Workload(const AdditionQueueDescriptor& descriptor, const WorkloadInfo& info);
+ NeonAdditionFloatWorkload(const AdditionQueueDescriptor& descriptor, const WorkloadInfo& info);
virtual void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.cpp
index ca5c8202cd..f24ff796cf 100644
--- a/src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonBatchNormalizationFloat32Workload.hpp"
+#include "NeonBatchNormalizationFloatWorkload.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
#include "../../../../include/armnn/ArmNN.hpp"
@@ -37,11 +37,11 @@ arm_compute::Status NeonBatchNormalizationValidate(const TensorInfo& input,
descriptor.m_Eps);
}
-NeonBatchNormalizationFloat32Workload::NeonBatchNormalizationFloat32Workload(
+NeonBatchNormalizationFloatWorkload::NeonBatchNormalizationFloatWorkload(
const BatchNormalizationQueueDescriptor& descriptor, const WorkloadInfo& info)
: FloatWorkload<BatchNormalizationQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("NeonBatchNormalizationFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonBatchNormalizationFloatWorkload", 1, 1);
arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -77,13 +77,13 @@ NeonBatchNormalizationFloat32Workload::NeonBatchNormalizationFloat32Workload(
FreeUnusedTensors();
}
-void NeonBatchNormalizationFloat32Workload::Execute() const
+void NeonBatchNormalizationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonBatchNormalizationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonBatchNormalizationFloatWorkload_Execute");
m_Layer.run();
}
-void NeonBatchNormalizationFloat32Workload::FreeUnusedTensors()
+void NeonBatchNormalizationFloatWorkload::FreeUnusedTensors()
{
FreeTensorIfUnused(m_Mean);
FreeTensorIfUnused(m_Variance);
diff --git a/src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.hpp
index 5eb5601f26..2844c650ca 100644
--- a/src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonBatchNormalizationFloatWorkload.hpp
@@ -18,11 +18,11 @@ arm_compute::Status NeonBatchNormalizationValidate(const TensorInfo& input,
const TensorInfo& gamma,
const BatchNormalizationDescriptor& descriptor);
-class NeonBatchNormalizationFloat32Workload : public FloatWorkload<BatchNormalizationQueueDescriptor>
+class NeonBatchNormalizationFloatWorkload : public FloatWorkload<BatchNormalizationQueueDescriptor>
{
public:
- NeonBatchNormalizationFloat32Workload(const BatchNormalizationQueueDescriptor& descriptor,
- const WorkloadInfo& info);
+ NeonBatchNormalizationFloatWorkload(const BatchNormalizationQueueDescriptor& descriptor,
+ const WorkloadInfo& info);
virtual void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonConstantFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonConstantFloatWorkload.cpp
index 4e5d570a8e..49749a8b08 100644
--- a/src/armnn/backends/NeonWorkloads/NeonConstantFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonConstantFloatWorkload.cpp
@@ -3,14 +3,14 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonConstantFloat32Workload.hpp"
+#include "NeonConstantFloatWorkload.hpp"
namespace armnn
{
-void NeonConstantFloat32Workload::Execute() const
+void NeonConstantFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonConstantFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonConstantFloatWorkload_Execute");
NeonBaseConstantWorkload::Execute();
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonConstantFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonConstantFloatWorkload.hpp
index 050954df24..8ae0add2c8 100644
--- a/src/armnn/backends/NeonWorkloads/NeonConstantFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonConstantFloatWorkload.hpp
@@ -10,7 +10,7 @@
namespace armnn
{
-class NeonConstantFloat32Workload : public NeonBaseConstantWorkload<DataType::Float16, DataType::Float32>
+class NeonConstantFloatWorkload : public NeonBaseConstantWorkload<DataType::Float16, DataType::Float32>
{
public:
using NeonBaseConstantWorkload<DataType::Float16, DataType::Float32>::NeonBaseConstantWorkload;
diff --git a/src/armnn/backends/NeonWorkloads/NeonConvolution2dFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonConvolution2dFloatWorkload.cpp
index 18ec6ca2e7..8037cb638f 100644
--- a/src/armnn/backends/NeonWorkloads/NeonConvolution2dFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonConvolution2dFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonConvolution2dFloat32Workload.hpp"
+#include "NeonConvolution2dFloatWorkload.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
#include "backends/NeonLayerSupport.hpp"
@@ -12,7 +12,7 @@ namespace armnn
{
using namespace armcomputetensorutils;
-NeonConvolution2dFloat32Workload::NeonConvolution2dFloat32Workload(const Convolution2dQueueDescriptor& descriptor,
+NeonConvolution2dFloatWorkload::NeonConvolution2dFloatWorkload(const Convolution2dQueueDescriptor& descriptor,
const WorkloadInfo& info, std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
: NeonConvolution2dBaseWorkload(descriptor, info, memoryManager)
{
@@ -25,15 +25,15 @@ NeonConvolution2dFloat32Workload::NeonConvolution2dFloat32Workload(const Convolu
FreeUnusedTensors();
}
-void NeonConvolution2dFloat32Workload::Execute() const
+void NeonConvolution2dFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonConvolution2dFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonConvolution2dFloatWorkload_Execute");
m_ConvolutionLayer->run();
}
-void NeonConvolution2dFloat32Workload::ValidateData() const
+void NeonConvolution2dFloatWorkload::ValidateData() const
{
- m_Data.ValidateInputsOutputs("NeonConvolution2dFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonConvolution2dFloatWorkload", 1, 1);
}
} //namespace armnn
diff --git a/src/armnn/backends/NeonWorkloads/NeonConvolution2dFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonConvolution2dFloatWorkload.hpp
index 0bb8d69d94..1d08118b61 100644
--- a/src/armnn/backends/NeonWorkloads/NeonConvolution2dFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonConvolution2dFloatWorkload.hpp
@@ -15,11 +15,11 @@
namespace armnn
{
-class NeonConvolution2dFloat32Workload : public NeonConvolution2dBaseWorkload<DataType::Float16, DataType::Float32>
+class NeonConvolution2dFloatWorkload : public NeonConvolution2dBaseWorkload<DataType::Float16, DataType::Float32>
{
public:
- NeonConvolution2dFloat32Workload(const Convolution2dQueueDescriptor& descriptor, const WorkloadInfo& info,
- std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
+ NeonConvolution2dFloatWorkload(const Convolution2dQueueDescriptor& descriptor, const WorkloadInfo& info,
+ std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
void Execute() const override;
void ValidateData() const override;
diff --git a/src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.cpp
index f94cd903b6..96ca938242 100644
--- a/src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonDepthwiseConvolutionFloat32Workload.hpp"
+#include "NeonDepthwiseConvolutionFloatWorkload.hpp"
#include "backends/NeonLayerSupport.hpp"
#include "backends/CpuTensorHandle.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
@@ -13,7 +13,7 @@ namespace armnn
{
using namespace armcomputetensorutils;
-NeonDepthwiseConvolutionFloat32Workload::NeonDepthwiseConvolutionFloat32Workload(
+NeonDepthwiseConvolutionFloatWorkload::NeonDepthwiseConvolutionFloatWorkload(
const DepthwiseConvolution2dQueueDescriptor& descriptor,
const WorkloadInfo& info)
: FloatWorkload<DepthwiseConvolution2dQueueDescriptor>(descriptor, info)
@@ -37,7 +37,7 @@ NeonDepthwiseConvolutionFloat32Workload::NeonDepthwiseConvolutionFloat32Workload
m_Data.m_Parameters.m_PadBottom,
arm_compute::DimensionRoundingType::FLOOR);
- m_Data.ValidateInputsOutputs("NeonDepthwiseConvolutionFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonDepthwiseConvolutionFloatWorkload", 1, 1);
arm_compute::ITensor& input = static_cast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ITensor& output = static_cast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -77,15 +77,15 @@ NeonDepthwiseConvolutionFloat32Workload::NeonDepthwiseConvolutionFloat32Workload
FreeUnusedTensors();
}
-void NeonDepthwiseConvolutionFloat32Workload::Execute() const
+void NeonDepthwiseConvolutionFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonDepthwiseConvolutionFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonDepthwiseConvolutionFloatWorkload_Execute");
BOOST_ASSERT(m_pDepthwiseConvolutionLayer);
m_pDepthwiseConvolutionLayer->run();
}
-void NeonDepthwiseConvolutionFloat32Workload::FreeUnusedTensors()
+void NeonDepthwiseConvolutionFloatWorkload::FreeUnusedTensors()
{
FreeTensorIfUnused(m_KernelTensor);
FreeTensorIfUnused(m_BiasTensor);
diff --git a/src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.hpp
index ece9f1877b..670182eef0 100644
--- a/src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonDepthwiseConvolutionFloatWorkload.hpp
@@ -10,11 +10,11 @@
namespace armnn
{
-class NeonDepthwiseConvolutionFloat32Workload : public FloatWorkload<DepthwiseConvolution2dQueueDescriptor>
+class NeonDepthwiseConvolutionFloatWorkload : public FloatWorkload<DepthwiseConvolution2dQueueDescriptor>
{
public:
- NeonDepthwiseConvolutionFloat32Workload(const DepthwiseConvolution2dQueueDescriptor& descriptor,
- const WorkloadInfo& info);
+ NeonDepthwiseConvolutionFloatWorkload(const DepthwiseConvolution2dQueueDescriptor& descriptor,
+ const WorkloadInfo& info);
virtual void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.cpp b/src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.cpp
new file mode 100644
index 0000000000..4d2cb029f5
--- /dev/null
+++ b/src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.cpp
@@ -0,0 +1,30 @@
+//
+// Copyright © 2017 Arm Ltd. All rights reserved.
+// See LICENSE file in the project root for full license information.
+//
+
+#include "NeonFloorFloatWorkload.hpp"
+
+namespace armnn
+{
+NeonFloorFloatWorkload::NeonFloorFloatWorkload(const FloorQueueDescriptor& descriptor,
+ const WorkloadInfo& info)
+ : FloatWorkload<FloorQueueDescriptor>(descriptor, info)
+{
+ m_Data.ValidateInputsOutputs("NeonFloorFloatWorkload", 1, 1);
+
+ arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
+ arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
+
+ m_Layer.configure(&input, &output);
+}
+
+void NeonFloorFloatWorkload::Execute() const
+{
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonFloorFloatWorkload_Execute");
+ m_Layer.run();
+}
+} //namespace armnn
+
+
+
diff --git a/src/armnn/backends/NeonWorkloads/NeonFloorFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.hpp
index 56680f1e39..1443d9a613 100644
--- a/src/armnn/backends/NeonWorkloads/NeonFloorFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonFloorFloatWorkload.hpp
@@ -10,10 +10,10 @@
namespace armnn
{
-class NeonFloorFloat32Workload : public FloatWorkload<FloorQueueDescriptor>
+class NeonFloorFloatWorkload : public FloatWorkload<FloorQueueDescriptor>
{
public:
- NeonFloorFloat32Workload(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info);
+ NeonFloorFloatWorkload(const FloorQueueDescriptor& descriptor, const WorkloadInfo& info);
virtual void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.cpp
index c3af41e20d..098d84bf0c 100644
--- a/src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonFullyConnectedFloat32Workload.hpp"
+#include "NeonFullyConnectedFloatWorkload.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
#include "backends/ArmComputeUtils.hpp"
@@ -42,12 +42,12 @@ arm_compute::Status NeonFullyConnectedWorkloadValidate(const TensorInfo& input,
fullyConnectedLayerInfo);
}
-NeonFullyConnectedFloat32Workload::NeonFullyConnectedFloat32Workload(const FullyConnectedQueueDescriptor& descriptor,
+NeonFullyConnectedFloatWorkload::NeonFullyConnectedFloatWorkload(const FullyConnectedQueueDescriptor& descriptor,
const WorkloadInfo& info, std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
: FloatWorkload<FullyConnectedQueueDescriptor>(descriptor, info)
, m_FullyConnectedLayer(memoryManager)
{
- m_Data.ValidateInputsOutputs("NeonFullyConnectedFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonFullyConnectedFloatWorkload", 1, 1);
arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -80,13 +80,13 @@ NeonFullyConnectedFloat32Workload::NeonFullyConnectedFloat32Workload(const Fully
FreeUnusedTensors();
}
-void NeonFullyConnectedFloat32Workload::Execute() const
+void NeonFullyConnectedFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonFullyConnectedFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonFullyConnectedFloatWorkload_Execute");
m_FullyConnectedLayer.run();
}
-void NeonFullyConnectedFloat32Workload::FreeUnusedTensors()
+void NeonFullyConnectedFloatWorkload::FreeUnusedTensors()
{
FreeTensorIfUnused(m_WeightsTensor);
FreeTensorIfUnused(m_BiasesTensor);
diff --git a/src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.hpp
index 684b5e0753..1f4464f0e1 100644
--- a/src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonFullyConnectedFloatWorkload.hpp
@@ -20,11 +20,11 @@ arm_compute::Status NeonFullyConnectedWorkloadValidate(const TensorInfo& input,
const TensorInfo& biases,
const FullyConnectedDescriptor& descriptor);
-class NeonFullyConnectedFloat32Workload : public FloatWorkload<FullyConnectedQueueDescriptor>
+class NeonFullyConnectedFloatWorkload : public FloatWorkload<FullyConnectedQueueDescriptor>
{
public:
- NeonFullyConnectedFloat32Workload(const FullyConnectedQueueDescriptor& descriptor, const WorkloadInfo& info,
- std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
+ NeonFullyConnectedFloatWorkload(const FullyConnectedQueueDescriptor& descriptor, const WorkloadInfo& info,
+ std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
virtual void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.cpp
index a3ae33f41f..bf0ef01349 100644
--- a/src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonL2NormalizationFloat32Workload.hpp"
+#include "NeonL2NormalizationFloatWorkload.hpp"
#include "backends/ArmComputeUtils.hpp"
namespace armnn
@@ -21,21 +21,21 @@ arm_compute::Status NeonL2NormalizationWorkloadValidate(const TensorInfo& input,
return arm_compute::NENormalizationLayer::validate(&aclInput, &aclOutput, normalizationInfo);
}
-NeonL2NormalizationFloat32Workload::NeonL2NormalizationFloat32Workload(const L2NormalizationQueueDescriptor& descriptor,
+NeonL2NormalizationFloatWorkload::NeonL2NormalizationFloatWorkload(const L2NormalizationQueueDescriptor& descriptor,
const WorkloadInfo& info, std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
: FloatWorkload<L2NormalizationQueueDescriptor>(descriptor, info)
, m_Layer(memoryManager)
{
- m_Data.ValidateInputsOutputs("NeonL2NormalizationFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonL2NormalizationFloatWorkload", 1, 1);
arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
m_Layer.configure(&input, &output, CreateAclNormalizationLayerInfoForL2Normalization(info.m_InputTensorInfos[0]));
}
-void NeonL2NormalizationFloat32Workload::Execute() const
+void NeonL2NormalizationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonL2NormalizationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonL2NormalizationFloatWorkload_Execute");
m_Layer.run();
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.hpp
index c3fcde5a57..5edff075f6 100644
--- a/src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonL2NormalizationFloatWorkload.hpp
@@ -17,11 +17,11 @@ namespace armnn
arm_compute::Status NeonL2NormalizationWorkloadValidate(const TensorInfo& input,
const TensorInfo& output);
-class NeonL2NormalizationFloat32Workload : public FloatWorkload<L2NormalizationQueueDescriptor>
+class NeonL2NormalizationFloatWorkload : public FloatWorkload<L2NormalizationQueueDescriptor>
{
public:
- NeonL2NormalizationFloat32Workload(const L2NormalizationQueueDescriptor& descriptor, const WorkloadInfo& info,
- std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
+ NeonL2NormalizationFloatWorkload(const L2NormalizationQueueDescriptor& descriptor, const WorkloadInfo& info,
+ std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
virtual void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonLstmFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonLstmFloatWorkload.cpp
index ba1369e179..8159e9feb4 100644
--- a/src/armnn/backends/NeonWorkloads/NeonLstmFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonLstmFloatWorkload.cpp
@@ -3,18 +3,18 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonLstmFloat32Workload.hpp"
+#include "NeonLstmFloatWorkload.hpp"
namespace armnn
{
-NeonLstmFloat32Workload::NeonLstmFloat32Workload(const LstmQueueDescriptor& descriptor,
- const WorkloadInfo& info)
+NeonLstmFloatWorkload::NeonLstmFloatWorkload(const LstmQueueDescriptor& descriptor,
+ const WorkloadInfo& info)
: FloatWorkload<LstmQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("NeonLstmFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonLstmFloatWorkload", 1, 1);
}
-void NeonLstmFloat32Workload::Execute() const
+void NeonLstmFloatWorkload::Execute() const
{
throw armnn::Exception("No implementation of Lstm in the Neon backend!");
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonLstmFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonLstmFloatWorkload.hpp
index 78ee1da341..fc06d55e6d 100644
--- a/src/armnn/backends/NeonWorkloads/NeonLstmFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonLstmFloatWorkload.hpp
@@ -10,10 +10,10 @@
namespace armnn
{
-class NeonLstmFloat32Workload : public FloatWorkload<LstmQueueDescriptor>
+class NeonLstmFloatWorkload : public FloatWorkload<LstmQueueDescriptor>
{
public:
- NeonLstmFloat32Workload(const LstmQueueDescriptor& descriptor, const WorkloadInfo& info);
+ NeonLstmFloatWorkload(const LstmQueueDescriptor& descriptor, const WorkloadInfo& info);
virtual void Execute() const override;
};
diff --git a/src/armnn/backends/NeonWorkloads/NeonMergerFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonMergerFloatWorkload.cpp
index 30dd283620..02caca91bc 100644
--- a/src/armnn/backends/NeonWorkloads/NeonMergerFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonMergerFloatWorkload.cpp
@@ -3,14 +3,14 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonMergerFloat32Workload.hpp"
+#include "NeonMergerFloatWorkload.hpp"
namespace armnn
{
-void NeonMergerFloat32Workload::Execute() const
+void NeonMergerFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonMergerFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonMergerFloatWorkload_Execute");
NeonBaseMergerWorkload::Execute();
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonMergerFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonMergerFloatWorkload.hpp
index 7b8ee9881f..c9f8e66e38 100644
--- a/src/armnn/backends/NeonWorkloads/NeonMergerFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonMergerFloatWorkload.hpp
@@ -10,7 +10,7 @@
namespace armnn
{
-class NeonMergerFloat32Workload : public NeonBaseMergerWorkload<DataType::Float16, DataType::Float32>
+class NeonMergerFloatWorkload : public NeonBaseMergerWorkload<DataType::Float16, DataType::Float32>
{
public:
using NeonBaseMergerWorkload<DataType::Float16, DataType::Float32>::NeonBaseMergerWorkload;
diff --git a/src/armnn/backends/NeonWorkloads/NeonMultiplicationFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonMultiplicationFloatWorkload.cpp
index a8a3cd77b4..ca38a24b09 100644
--- a/src/armnn/backends/NeonWorkloads/NeonMultiplicationFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonMultiplicationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonMultiplicationFloat32Workload.hpp"
+#include "NeonMultiplicationFloatWorkload.hpp"
namespace armnn
@@ -28,11 +28,11 @@ arm_compute::Status NeonMultiplicationWorkloadValidate(const TensorInfo& input0,
arm_compute::RoundingPolicy::TO_ZERO);
}
-NeonMultiplicationFloat32Workload::NeonMultiplicationFloat32Workload(const MultiplicationQueueDescriptor& descriptor,
- const WorkloadInfo& info)
+NeonMultiplicationFloatWorkload::NeonMultiplicationFloatWorkload(const MultiplicationQueueDescriptor& descriptor,
+ const WorkloadInfo& info)
: FloatWorkload<MultiplicationQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("NeonMultiplicationFloat32Workload", 2, 1);
+ m_Data.ValidateInputsOutputs("NeonMultiplicationFloatWorkload", 2, 1);
arm_compute::ITensor& input1 = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ITensor& input2 = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[1])->GetTensor();
@@ -49,9 +49,9 @@ NeonMultiplicationFloat32Workload::NeonMultiplicationFloat32Workload(const Multi
arm_compute::RoundingPolicy::TO_ZERO);
}
-void NeonMultiplicationFloat32Workload::Execute() const
+void NeonMultiplicationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonMultiplicationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonMultiplicationFloatWorkload_Execute");
m_PixelWiseMultiplication.run();
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonMultiplicationFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonMultiplicationFloatWorkload.hpp
index 62e84a2e07..eb0f7741f0 100644
--- a/src/armnn/backends/NeonWorkloads/NeonMultiplicationFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonMultiplicationFloatWorkload.hpp
@@ -13,10 +13,10 @@ arm_compute::Status NeonMultiplicationWorkloadValidate(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output);
-class NeonMultiplicationFloat32Workload : public FloatWorkload<MultiplicationQueueDescriptor>
+class NeonMultiplicationFloatWorkload : public FloatWorkload<MultiplicationQueueDescriptor>
{
public:
- NeonMultiplicationFloat32Workload(const MultiplicationQueueDescriptor& descriptor, const WorkloadInfo& info);
+ NeonMultiplicationFloatWorkload(const MultiplicationQueueDescriptor& descriptor, const WorkloadInfo& info);
virtual void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonNormalizationFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonNormalizationFloatWorkload.cpp
index 20936a2760..8c2a87d8bc 100644
--- a/src/armnn/backends/NeonWorkloads/NeonNormalizationFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonNormalizationFloatWorkload.cpp
@@ -3,7 +3,7 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonNormalizationFloat32Workload.hpp"
+#include "NeonNormalizationFloatWorkload.hpp"
#include "backends/NeonLayerSupport.hpp"
#include "backends/ArmComputeUtils.hpp"
#include "backends/ArmComputeTensorUtils.hpp"
@@ -24,13 +24,13 @@ arm_compute::Status NeonNormalizationWorkloadValidate(const TensorInfo& input,
return arm_compute::NENormalizationLayer::validate(&aclInput, &aclOutput, normalizationInfo);
}
-NeonNormalizationFloat32Workload::NeonNormalizationFloat32Workload(const NormalizationQueueDescriptor& descriptor,
+NeonNormalizationFloatWorkload::NeonNormalizationFloatWorkload(const NormalizationQueueDescriptor& descriptor,
const WorkloadInfo& info,
std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
: FloatWorkload<NormalizationQueueDescriptor>(descriptor, info)
, m_NormalizationLayer(memoryManager)
{
- m_Data.ValidateInputsOutputs("NeonNormalizationFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonNormalizationFloatWorkload", 1, 1);
std::string reasonIfUnsupported;
if (!IsNeonNormalizationDescParamsSupported(&reasonIfUnsupported, m_Data.m_Parameters))
{
@@ -61,9 +61,9 @@ NeonNormalizationFloat32Workload::NeonNormalizationFloat32Workload(const Normali
m_NormalizationLayer.configure(&input, &output, normalizationInfo);
}
-void NeonNormalizationFloat32Workload::Execute() const
+void NeonNormalizationFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonNormalizationFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonNormalizationFloatWorkload_Execute");
m_NormalizationLayer.run();
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonNormalizationFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonNormalizationFloatWorkload.hpp
index 8f0823454b..8e8cc40c5a 100644
--- a/src/armnn/backends/NeonWorkloads/NeonNormalizationFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonNormalizationFloatWorkload.hpp
@@ -16,11 +16,11 @@ arm_compute::Status NeonNormalizationWorkloadValidate(const TensorInfo& input,
const TensorInfo& output,
const NormalizationDescriptor& descriptor);
-class NeonNormalizationFloat32Workload : public FloatWorkload<NormalizationQueueDescriptor>
+class NeonNormalizationFloatWorkload : public FloatWorkload<NormalizationQueueDescriptor>
{
public:
- NeonNormalizationFloat32Workload(const NormalizationQueueDescriptor& descriptor, const WorkloadInfo& info,
- std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
+ NeonNormalizationFloatWorkload(const NormalizationQueueDescriptor& descriptor, const WorkloadInfo& info,
+ std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
virtual void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.cpp
deleted file mode 100644
index cb690c51b8..0000000000
--- a/src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.cpp
+++ /dev/null
@@ -1,27 +0,0 @@
-//
-// Copyright © 2017 Arm Ltd. All rights reserved.
-// See LICENSE file in the project root for full license information.
-//
-
-#include "NeonPooling2dFloat32Workload.hpp"
-
-
-
-namespace armnn
-{
-
-NeonPooling2dFloat32Workload::NeonPooling2dFloat32Workload(const Pooling2dQueueDescriptor& descriptor,
- const WorkloadInfo& info)
- : NeonPooling2dBaseWorkload<armnn::DataType::Float16, armnn::DataType::Float32>(descriptor, info,
- "NeonPooling2dFloat32Workload")
-{
-}
-
-void NeonPooling2dFloat32Workload::Execute() const
-{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonPooling2dFloat32Workload_Execute");
- m_PoolingLayer.run();
-}
-
-} //namespace armnn
-
diff --git a/src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.cpp b/src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.cpp
new file mode 100644
index 0000000000..90e09dbaf4
--- /dev/null
+++ b/src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.cpp
@@ -0,0 +1,27 @@
+//
+// Copyright © 2017 Arm Ltd. All rights reserved.
+// See LICENSE file in the project root for full license information.
+//
+
+#include "NeonPooling2dFloatWorkload.hpp"
+
+
+
+namespace armnn
+{
+
+NeonPooling2dFloatWorkload::NeonPooling2dFloatWorkload(const Pooling2dQueueDescriptor& descriptor,
+ const WorkloadInfo& info)
+ : NeonPooling2dBaseWorkload<armnn::DataType::Float16, armnn::DataType::Float32>(descriptor, info,
+ "NeonPooling2dFloatWorkload")
+{
+}
+
+void NeonPooling2dFloatWorkload::Execute() const
+{
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonPooling2dFloatWorkload_Execute");
+ m_PoolingLayer.run();
+}
+
+} //namespace armnn
+
diff --git a/src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.hpp
index 36c4e7edf1..ba4e9cab6d 100644
--- a/src/armnn/backends/NeonWorkloads/NeonPooling2dFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonPooling2dFloatWorkload.hpp
@@ -11,11 +11,11 @@
namespace armnn
{
-class NeonPooling2dFloat32Workload : public NeonPooling2dBaseWorkload<armnn::DataType::Float16,
- armnn::DataType::Float32>
+class NeonPooling2dFloatWorkload : public NeonPooling2dBaseWorkload<armnn::DataType::Float16,
+ armnn::DataType::Float32>
{
public:
- NeonPooling2dFloat32Workload(const Pooling2dQueueDescriptor& descriptor, const WorkloadInfo& info);
+ NeonPooling2dFloatWorkload(const Pooling2dQueueDescriptor& descriptor, const WorkloadInfo& info);
virtual void Execute() const override;
};
diff --git a/src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.cpp
deleted file mode 100644
index 93f6eb8ef5..0000000000
--- a/src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.cpp
+++ /dev/null
@@ -1,32 +0,0 @@
-//
-// Copyright © 2017 Arm Ltd. All rights reserved.
-// See LICENSE file in the project root for full license information.
-//
-
-#include "NeonReshapeFloat32Workload.hpp"
-
-
-
-namespace armnn
-{
-
-NeonReshapeFloat32Workload::NeonReshapeFloat32Workload(const ReshapeQueueDescriptor& descriptor,
- const WorkloadInfo& info)
- : FloatWorkload<ReshapeQueueDescriptor>(descriptor, info)
-{
- m_Data.ValidateInputsOutputs("NeonReshapeFloat32Workload", 1, 1);
-
- arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
- arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
-
- m_Layer.configure(&input, &output);
-}
-
-void NeonReshapeFloat32Workload::Execute() const
-{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonReshapeFloat32Workload_Execute");
- m_Layer.run();
-}
-
-} //namespace armnn
-
diff --git a/src/armnn/backends/NeonWorkloads/NeonFloorFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonReshapeFloatWorkload.cpp
index c43cfa9c46..ef229b252f 100644
--- a/src/armnn/backends/NeonWorkloads/NeonFloorFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonReshapeFloatWorkload.cpp
@@ -3,15 +3,18 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonFloorFloat32Workload.hpp"
+#include "NeonReshapeFloatWorkload.hpp"
+
+
namespace armnn
{
-NeonFloorFloat32Workload::NeonFloorFloat32Workload(const FloorQueueDescriptor& descriptor,
+
+NeonReshapeFloatWorkload::NeonReshapeFloatWorkload(const ReshapeQueueDescriptor& descriptor,
const WorkloadInfo& info)
- : FloatWorkload<FloorQueueDescriptor>(descriptor, info)
+ : FloatWorkload<ReshapeQueueDescriptor>(descriptor, info)
{
- m_Data.ValidateInputsOutputs("NeonFloorFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonReshapeFloatWorkload", 1, 1);
arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
arm_compute::ITensor& output = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Outputs[0])->GetTensor();
@@ -19,12 +22,11 @@ NeonFloorFloat32Workload::NeonFloorFloat32Workload(const FloorQueueDescriptor& d
m_Layer.configure(&input, &output);
}
-void NeonFloorFloat32Workload::Execute() const
+void NeonReshapeFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonFloorFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonReshapeFloatWorkload_Execute");
m_Layer.run();
}
-} //namespace armnn
-
+} //namespace armnn
diff --git a/src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonReshapeFloatWorkload.hpp
index 3e5cca1b9e..d32d779962 100644
--- a/src/armnn/backends/NeonWorkloads/NeonReshapeFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonReshapeFloatWorkload.hpp
@@ -10,10 +10,10 @@
namespace armnn
{
-class NeonReshapeFloat32Workload : public FloatWorkload<ReshapeQueueDescriptor>
+class NeonReshapeFloatWorkload : public FloatWorkload<ReshapeQueueDescriptor>
{
public:
- NeonReshapeFloat32Workload(const ReshapeQueueDescriptor& descriptor, const WorkloadInfo& info);
+ NeonReshapeFloatWorkload(const ReshapeQueueDescriptor& descriptor, const WorkloadInfo& info);
virtual void Execute() const override;
diff --git a/src/armnn/backends/NeonWorkloads/NeonSoftmaxFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonSoftmaxFloatWorkload.cpp
index 027b508ad5..981e63ea55 100644
--- a/src/armnn/backends/NeonWorkloads/NeonSoftmaxFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonSoftmaxFloatWorkload.cpp
@@ -3,17 +3,17 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonSoftmaxFloat32Workload.hpp"
+#include "NeonSoftmaxFloatWorkload.hpp"
namespace armnn
{
-NeonSoftmaxFloat32Workload::NeonSoftmaxFloat32Workload(const SoftmaxQueueDescriptor& descriptor,
+NeonSoftmaxFloatWorkload::NeonSoftmaxFloatWorkload(const SoftmaxQueueDescriptor& descriptor,
const WorkloadInfo& info, std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager)
: FloatWorkload<SoftmaxQueueDescriptor>(descriptor, info)
, m_SoftmaxLayer(memoryManager)
{
- m_Data.ValidateInputsOutputs("NeonSoftmaxFloat32Workload", 1, 1);
+ m_Data.ValidateInputsOutputs("NeonSoftmaxFloatWorkload", 1, 1);
// The ArmCompute softmax layer uses 2D input/output tensors, so flatten the first three dimensions.
arm_compute::ITensor& input = boost::polymorphic_downcast<INeonTensorHandle*>(m_Data.m_Inputs[0])->GetTensor();
@@ -22,9 +22,9 @@ NeonSoftmaxFloat32Workload::NeonSoftmaxFloat32Workload(const SoftmaxQueueDescrip
m_SoftmaxLayer.configure(&input, &output, m_Data.m_Parameters.m_Beta);
}
-void NeonSoftmaxFloat32Workload::Execute() const
+void NeonSoftmaxFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonSoftmaxFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonSoftmaxFloatWorkload_Execute");
m_SoftmaxLayer.run();
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonSoftmaxFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonSoftmaxFloatWorkload.hpp
index 3656a26a3c..52dbded5dd 100644
--- a/src/armnn/backends/NeonWorkloads/NeonSoftmaxFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonSoftmaxFloatWorkload.hpp
@@ -14,11 +14,11 @@
namespace armnn
{
-class NeonSoftmaxFloat32Workload : public FloatWorkload<SoftmaxQueueDescriptor>
+class NeonSoftmaxFloatWorkload : public FloatWorkload<SoftmaxQueueDescriptor>
{
public:
- NeonSoftmaxFloat32Workload(const SoftmaxQueueDescriptor& descriptor, const WorkloadInfo& info,
- std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
+ NeonSoftmaxFloatWorkload(const SoftmaxQueueDescriptor& descriptor, const WorkloadInfo& info,
+ std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager);
virtual void Execute() const override;
private:
diff --git a/src/armnn/backends/NeonWorkloads/NeonSplitterFloat32Workload.cpp b/src/armnn/backends/NeonWorkloads/NeonSplitterFloatWorkload.cpp
index 996fc15adb..06e300cd5a 100644
--- a/src/armnn/backends/NeonWorkloads/NeonSplitterFloat32Workload.cpp
+++ b/src/armnn/backends/NeonWorkloads/NeonSplitterFloatWorkload.cpp
@@ -3,14 +3,14 @@
// See LICENSE file in the project root for full license information.
//
-#include "NeonSplitterFloat32Workload.hpp"
+#include "NeonSplitterFloatWorkload.hpp"
namespace armnn
{
-void NeonSplitterFloat32Workload::Execute() const
+void NeonSplitterFloatWorkload::Execute() const
{
- ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonSplitterFloat32Workload_Execute");
+ ARMNN_SCOPED_PROFILING_EVENT_NEON("NeonSplitterFloatWorkload_Execute");
NeonBaseSplitterWorkload::Execute();
}
diff --git a/src/armnn/backends/NeonWorkloads/NeonSplitterFloat32Workload.hpp b/src/armnn/backends/NeonWorkloads/NeonSplitterFloatWorkload.hpp
index 9f6dc75499..5d6b328413 100644
--- a/src/armnn/backends/NeonWorkloads/NeonSplitterFloat32Workload.hpp
+++ b/src/armnn/backends/NeonWorkloads/NeonSplitterFloatWorkload.hpp
@@ -10,7 +10,7 @@
namespace armnn
{
-class NeonSplitterFloat32Workload : public NeonBaseSplitterWorkload<DataType::Float16, DataType::Float32>
+class NeonSplitterFloatWorkload : public NeonBaseSplitterWorkload<DataType::Float16, DataType::Float32>
{
public:
using NeonBaseSplitterWorkload<DataType::Float16, DataType::Float32>::NeonBaseSplitterWorkload;
diff --git a/src/armnn/backends/test/CreateWorkloadCl.cpp b/src/armnn/backends/test/CreateWorkloadCl.cpp
index 5d4265911f..538d5af667 100644
--- a/src/armnn/backends/test/CreateWorkloadCl.cpp
+++ b/src/armnn/backends/test/CreateWorkloadCl.cpp
@@ -37,14 +37,14 @@ static void ClCreateActivationWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, {1}));
}
-BOOST_AUTO_TEST_CASE(CreateActivationFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateActivationFloatWorkload)
{
- ClCreateActivationWorkloadTest<ClActivationFloat32Workload, armnn::DataType::Float32>();
+ ClCreateActivationWorkloadTest<ClActivationFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateActivationFloat16Workload)
{
- ClCreateActivationWorkloadTest<ClActivationFloat32Workload, armnn::DataType::Float16>();
+ ClCreateActivationWorkloadTest<ClActivationFloatWorkload, armnn::DataType::Float16>();
}
template <typename AdditionWorkloadType, armnn::DataType DataType>
@@ -64,14 +64,14 @@ static void ClCreateAdditionWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, {2, 3}));
}
-BOOST_AUTO_TEST_CASE(CreateAdditionFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateAdditionFloatWorkload)
{
- ClCreateAdditionWorkloadTest<ClAdditionFloat32Workload, armnn::DataType::Float32>();
+ ClCreateAdditionWorkloadTest<ClAdditionFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateAdditionFloat16Workload)
{
- ClCreateAdditionWorkloadTest<ClAdditionFloat32Workload, armnn::DataType::Float16>();
+ ClCreateAdditionWorkloadTest<ClAdditionFloatWorkload, armnn::DataType::Float16>();
}
template <typename BatchNormalizationWorkloadType, armnn::DataType DataType>
@@ -92,14 +92,14 @@ static void ClCreateBatchNormalizationWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, {2, 3, 1, 1}));
}
-BOOST_AUTO_TEST_CASE(CreateBatchNormalizationFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateBatchNormalizationFloatWorkload)
{
- ClCreateBatchNormalizationWorkloadTest<ClBatchNormalizationFloat32Workload, armnn::DataType::Float32>();
+ ClCreateBatchNormalizationWorkloadTest<ClBatchNormalizationFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateBatchNormalizationFloat16Workload)
{
- ClCreateBatchNormalizationWorkloadTest<ClBatchNormalizationFloat32Workload, armnn::DataType::Float16>();
+ ClCreateBatchNormalizationWorkloadTest<ClBatchNormalizationFloatWorkload, armnn::DataType::Float16>();
}
BOOST_AUTO_TEST_CASE(CreateConvertFp16ToFp32Workload)
@@ -150,14 +150,14 @@ static void ClConvolution2dWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, {2, 2, 2, 10}));
}
-BOOST_AUTO_TEST_CASE(CreateConvolution2dFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateConvolution2dFloatWorkload)
{
- ClConvolution2dWorkloadTest<ClConvolution2dFloat32Workload, armnn::DataType::Float32>();
+ ClConvolution2dWorkloadTest<ClConvolution2dFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateConvolution2dFloat16Workload)
{
- ClConvolution2dWorkloadTest<ClConvolution2dFloat32Workload, armnn::DataType::Float16>();
+ ClConvolution2dWorkloadTest<ClConvolution2dFloatWorkload, armnn::DataType::Float16>();
}
@@ -177,14 +177,14 @@ static void ClDirectConvolution2dWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, {2, 2, 6, 6}));
}
-BOOST_AUTO_TEST_CASE(CreateDirectConvolution2dFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateDirectConvolution2dFloatWorkload)
{
- ClDirectConvolution2dWorkloadTest<ClConvolution2dFloat32Workload, armnn::DataType::Float32>();
+ ClDirectConvolution2dWorkloadTest<ClConvolution2dFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateDirectConvolution2dFloat16Workload)
{
- ClDirectConvolution2dWorkloadTest<ClConvolution2dFloat32Workload, armnn::DataType::Float16>();
+ ClDirectConvolution2dWorkloadTest<ClConvolution2dFloatWorkload, armnn::DataType::Float16>();
}
BOOST_AUTO_TEST_CASE(CreateDirectConvolution2dUint8Workload)
@@ -209,14 +209,14 @@ static void ClCreateFullyConnectedWorkloadTest()
}
-BOOST_AUTO_TEST_CASE(CreateFullyConnectedFloat32WorkloadTest)
+BOOST_AUTO_TEST_CASE(CreateFullyConnectedFloatWorkloadTest)
{
- ClCreateFullyConnectedWorkloadTest<ClFullyConnectedFloat32Workload, armnn::DataType::Float32>();
+ ClCreateFullyConnectedWorkloadTest<ClFullyConnectedFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateFullyConnectedFloat16WorkloadTest)
{
- ClCreateFullyConnectedWorkloadTest<ClFullyConnectedFloat32Workload, armnn::DataType::Float16>();
+ ClCreateFullyConnectedWorkloadTest<ClFullyConnectedFloatWorkload, armnn::DataType::Float16>();
}
@@ -239,14 +239,14 @@ static void ClCreateMultiplicationWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, {2, 3}));
}
-BOOST_AUTO_TEST_CASE(CreateMultiplicationFloat32WorkloadTest)
+BOOST_AUTO_TEST_CASE(CreateMultiplicationFloatWorkloadTest)
{
- ClCreateMultiplicationWorkloadTest<ClMultiplicationFloat32Workload, armnn::DataType::Float32>();
+ ClCreateMultiplicationWorkloadTest<ClMultiplicationFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateMultiplicationFloat16WorkloadTest)
{
- ClCreateMultiplicationWorkloadTest<ClMultiplicationFloat32Workload, armnn::DataType::Float16>();
+ ClCreateMultiplicationWorkloadTest<ClMultiplicationFloatWorkload, armnn::DataType::Float16>();
}
template <typename NormalizationWorkloadType, typename armnn::DataType DataType>
@@ -267,14 +267,14 @@ static void ClNormalizationWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, {3, 5, 5, 1}));
}
-BOOST_AUTO_TEST_CASE(CreateNormalizationFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateNormalizationFloatWorkload)
{
- ClNormalizationWorkloadTest<ClNormalizationFloat32Workload, armnn::DataType::Float32>();
+ ClNormalizationWorkloadTest<ClNormalizationFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateNormalizationFloat16Workload)
{
- ClNormalizationWorkloadTest<ClNormalizationFloat32Workload, armnn::DataType::Float16>();
+ ClNormalizationWorkloadTest<ClNormalizationFloatWorkload, armnn::DataType::Float16>();
}
template <typename Pooling2dWorkloadType, typename armnn::DataType DataType>
@@ -294,14 +294,14 @@ static void ClPooling2dWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, {3, 2, 2, 4}));
}
-BOOST_AUTO_TEST_CASE(CreatePooling2dFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreatePooling2dFloatWorkload)
{
- ClPooling2dWorkloadTest<ClPooling2dFloat32Workload, armnn::DataType::Float32>();
+ ClPooling2dWorkloadTest<ClPooling2dFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreatePooling2dFloat16Workload)
{
- ClPooling2dWorkloadTest<ClPooling2dFloat32Workload, armnn::DataType::Float16>();
+ ClPooling2dWorkloadTest<ClPooling2dFloatWorkload, armnn::DataType::Float16>();
}
template <typename ReshapeWorkloadType, typename armnn::DataType DataType>
@@ -321,14 +321,14 @@ static void ClCreateReshapeWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, {4})); // Leading size 1 dimensions are collapsed by ACL.
}
-BOOST_AUTO_TEST_CASE(CreateReshapeFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateReshapeFloatWorkload)
{
- ClCreateReshapeWorkloadTest<ClReshapeFloat32Workload, armnn::DataType::Float32>();
+ ClCreateReshapeWorkloadTest<ClReshapeFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateReshapeFloat16Workload)
{
- ClCreateReshapeWorkloadTest<ClReshapeFloat32Workload, armnn::DataType::Float16>();
+ ClCreateReshapeWorkloadTest<ClReshapeFloatWorkload, armnn::DataType::Float16>();
}
BOOST_AUTO_TEST_CASE(CreateReshapeUint8Workload)
@@ -344,7 +344,7 @@ static void ClSoftmaxWorkloadTest()
auto workload = CreateSoftmaxWorkloadTest<SoftmaxWorkloadType, DataType>(factory, graph);
- // Checks that inputs/outputs are as we expect them (see definition of ClSoftmaxFloat32Workload).
+ // Checks that inputs/outputs are as we expect them (see definition of ClSoftmaxFloatWorkload).
SoftmaxQueueDescriptor queueDescriptor = workload->GetData();
auto inputHandle = boost::polymorphic_downcast<IClTensorHandle*>(queueDescriptor.m_Inputs[0]);
auto outputHandle = boost::polymorphic_downcast<IClTensorHandle*>(queueDescriptor.m_Outputs[0]);
@@ -354,14 +354,14 @@ static void ClSoftmaxWorkloadTest()
}
-BOOST_AUTO_TEST_CASE(CreateSoftmaxFloat32WorkloadTest)
+BOOST_AUTO_TEST_CASE(CreateSoftmaxFloatWorkloadTest)
{
- ClSoftmaxWorkloadTest<ClSoftmaxFloat32Workload, armnn::DataType::Float32>();
+ ClSoftmaxWorkloadTest<ClSoftmaxFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateSoftmaxFloat16WorkloadTest)
{
- ClSoftmaxWorkloadTest<ClSoftmaxFloat32Workload, armnn::DataType::Float16>();
+ ClSoftmaxWorkloadTest<ClSoftmaxFloatWorkload, armnn::DataType::Float16>();
}
template <typename SplitterWorkloadType, typename armnn::DataType DataType>
@@ -389,14 +389,14 @@ static void ClSplitterWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle0, {7, 7}));
}
-BOOST_AUTO_TEST_CASE(CreateSplitterFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateSplitterFloatWorkload)
{
- ClSplitterWorkloadTest<ClSplitterFloat32Workload, armnn::DataType::Float32>();
+ ClSplitterWorkloadTest<ClSplitterFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateSplitterFloat16Workload)
{
- ClSplitterWorkloadTest<ClSplitterFloat32Workload, armnn::DataType::Float16>();
+ ClSplitterWorkloadTest<ClSplitterFloatWorkload, armnn::DataType::Float16>();
}
template <typename SplitterWorkloadType, typename MergerWorkloadType, typename armnn::DataType DataType>
@@ -441,14 +441,14 @@ static void ClSplitterMergerTest()
BOOST_TEST(validSubTensorParents);
}
-BOOST_AUTO_TEST_CASE(CreateSplitterMergerFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateSplitterMergerFloatWorkload)
{
- ClSplitterMergerTest<ClSplitterFloat32Workload, ClMergerFloat32Workload, armnn::DataType::Float32>();
+ ClSplitterMergerTest<ClSplitterFloatWorkload, ClMergerFloatWorkload, armnn::DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateSplitterMergerFloat16Workload)
{
- ClSplitterMergerTest<ClSplitterFloat32Workload, ClMergerFloat32Workload, armnn::DataType::Float16>();
+ ClSplitterMergerTest<ClSplitterFloatWorkload, ClMergerFloatWorkload, armnn::DataType::Float16>();
}
@@ -459,14 +459,14 @@ BOOST_AUTO_TEST_CASE(CreateSingleOutputMultipleInputs)
Graph graph;
ClWorkloadFactory factory;
- std::unique_ptr<ClSplitterFloat32Workload> wlSplitter;
- std::unique_ptr<ClActivationFloat32Workload> wlActiv0_0;
- std::unique_ptr<ClActivationFloat32Workload> wlActiv0_1;
- std::unique_ptr<ClActivationFloat32Workload> wlActiv1_0;
- std::unique_ptr<ClActivationFloat32Workload> wlActiv1_1;
-
- CreateSplitterMultipleInputsOneOutputWorkloadTest<ClSplitterFloat32Workload,
- ClActivationFloat32Workload, armnn::DataType::Float32>(factory, graph, wlSplitter, wlActiv0_0, wlActiv0_1,
+ std::unique_ptr<ClSplitterFloatWorkload> wlSplitter;
+ std::unique_ptr<ClActivationFloatWorkload> wlActiv0_0;
+ std::unique_ptr<ClActivationFloatWorkload> wlActiv0_1;
+ std::unique_ptr<ClActivationFloatWorkload> wlActiv1_0;
+ std::unique_ptr<ClActivationFloatWorkload> wlActiv1_1;
+
+ CreateSplitterMultipleInputsOneOutputWorkloadTest<ClSplitterFloatWorkload,
+ ClActivationFloatWorkload, armnn::DataType::Float32>(factory, graph, wlSplitter, wlActiv0_0, wlActiv0_1,
wlActiv1_0, wlActiv1_1);
//Checks that the index of inputs/outputs matches what we declared on InputDescriptor construction.
@@ -502,7 +502,7 @@ BOOST_AUTO_TEST_CASE(CreateL2NormalizationWorkload)
Graph graph;
ClWorkloadFactory factory;
- auto workload = CreateL2NormalizationWorkloadTest<ClL2NormalizationFloat32Workload, armnn::DataType::Float32>
+ auto workload = CreateL2NormalizationWorkloadTest<ClL2NormalizationFloatWorkload, armnn::DataType::Float32>
(factory, graph);
// Checks that inputs/outputs are as we expect them (see definition of CreateNormalizationWorkloadTest).
@@ -528,9 +528,9 @@ static void ClCreateLstmWorkloadTest()
BOOST_TEST(CompareIClTensorHandleShape(outputHandle, { 2, 4 }));
}
-BOOST_AUTO_TEST_CASE(CreateLSTMWorkloadFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateLSTMWorkloadFloatWorkload)
{
- ClCreateLstmWorkloadTest<ClLstmFloat32Workload>();
+ ClCreateLstmWorkloadTest<ClLstmFloatWorkload>();
}
diff --git a/src/armnn/backends/test/CreateWorkloadNeon.cpp b/src/armnn/backends/test/CreateWorkloadNeon.cpp
index b2a444af74..d84b39b339 100644
--- a/src/armnn/backends/test/CreateWorkloadNeon.cpp
+++ b/src/armnn/backends/test/CreateWorkloadNeon.cpp
@@ -69,13 +69,13 @@ static void NeonCreateActivationWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreateActivationFloat16Workload)
{
- NeonCreateActivationWorkloadTest<NeonActivationFloat32Workload, DataType::Float16>();
+ NeonCreateActivationWorkloadTest<NeonActivationFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreateActivationFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateActivationFloatWorkload)
{
- NeonCreateActivationWorkloadTest<NeonActivationFloat32Workload, DataType::Float32>();
+ NeonCreateActivationWorkloadTest<NeonActivationFloatWorkload, DataType::Float32>();
}
template <typename AdditionWorkloadType, typename armnn::DataType DataType>
@@ -98,13 +98,13 @@ static void NeonCreateAdditionWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreateAdditionFloat16Workload)
{
- NeonCreateAdditionWorkloadTest<NeonAdditionFloat32Workload, DataType::Float16>();
+ NeonCreateAdditionWorkloadTest<NeonAdditionFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreateAdditionFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateAdditionFloatWorkload)
{
- NeonCreateAdditionWorkloadTest<NeonAdditionFloat32Workload, DataType::Float32>();
+ NeonCreateAdditionWorkloadTest<NeonAdditionFloatWorkload, DataType::Float32>();
}
template <typename BatchNormalizationWorkloadType, typename armnn::DataType DataType>
@@ -125,13 +125,13 @@ static void NeonCreateBatchNormalizationWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreateBatchNormalizationFloat16Workload)
{
- NeonCreateBatchNormalizationWorkloadTest<NeonBatchNormalizationFloat32Workload, DataType::Float16>();
+ NeonCreateBatchNormalizationWorkloadTest<NeonBatchNormalizationFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreateBatchNormalizationFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateBatchNormalizationFloatWorkload)
{
- NeonCreateBatchNormalizationWorkloadTest<NeonBatchNormalizationFloat32Workload, DataType::Float32>();
+ NeonCreateBatchNormalizationWorkloadTest<NeonBatchNormalizationFloatWorkload, DataType::Float32>();
}
template <typename Convolution2dWorkloadType, typename armnn::DataType DataType>
@@ -153,13 +153,13 @@ static void NeonCreateConvolution2dWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreateConvolution2dFloat16Workload)
{
- NeonCreateConvolution2dWorkloadTest<NeonConvolution2dFloat32Workload, DataType::Float16>();
+ NeonCreateConvolution2dWorkloadTest<NeonConvolution2dFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreateConvolution2dFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateConvolution2dFloatWorkload)
{
- NeonCreateConvolution2dWorkloadTest<NeonConvolution2dFloat32Workload, DataType::Float32>();
+ NeonCreateConvolution2dWorkloadTest<NeonConvolution2dFloatWorkload, DataType::Float32>();
}
template <typename FullyConnectedWorkloadType, typename armnn::DataType DataType>
@@ -181,13 +181,13 @@ static void NeonCreateFullyConnectedWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreateFullyConnectedFloat16Workload)
{
- NeonCreateFullyConnectedWorkloadTest<NeonFullyConnectedFloat32Workload, DataType::Float16>();
+ NeonCreateFullyConnectedWorkloadTest<NeonFullyConnectedFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreateFullyConnectedFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateFullyConnectedFloatWorkload)
{
- NeonCreateFullyConnectedWorkloadTest<NeonFullyConnectedFloat32Workload, DataType::Float32>();
+ NeonCreateFullyConnectedWorkloadTest<NeonFullyConnectedFloatWorkload, DataType::Float32>();
}
template <typename MultiplicationWorkloadType, typename armnn::DataType DataType>
@@ -211,13 +211,13 @@ static void NeonCreateMultiplicationWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreateMultiplicationFloat16Workload)
{
- NeonCreateMultiplicationWorkloadTest<NeonMultiplicationFloat32Workload, DataType::Float16>();
+ NeonCreateMultiplicationWorkloadTest<NeonMultiplicationFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreateMultiplicationFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateMultiplicationFloatWorkload)
{
- NeonCreateMultiplicationWorkloadTest<NeonMultiplicationFloat32Workload, DataType::Float32>();
+ NeonCreateMultiplicationWorkloadTest<NeonMultiplicationFloatWorkload, DataType::Float32>();
}
template <typename NormalizationWorkloadType, typename armnn::DataType DataType>
@@ -238,13 +238,13 @@ static void NeonCreateNormalizationWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreateNormalizationFloat16Workload)
{
- NeonCreateNormalizationWorkloadTest<NeonNormalizationFloat32Workload, DataType::Float16>();
+ NeonCreateNormalizationWorkloadTest<NeonNormalizationFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreateNormalizationFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateNormalizationFloatWorkload)
{
- NeonCreateNormalizationWorkloadTest<NeonNormalizationFloat32Workload, DataType::Float32>();
+ NeonCreateNormalizationWorkloadTest<NeonNormalizationFloatWorkload, DataType::Float32>();
}
template <typename Pooling2dWorkloadType, typename armnn::DataType DataType>
@@ -266,13 +266,13 @@ static void NeonCreatePooling2dWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreatePooling2dFloat16Workload)
{
- NeonCreatePooling2dWorkloadTest<NeonPooling2dFloat32Workload, DataType::Float16>();
+ NeonCreatePooling2dWorkloadTest<NeonPooling2dFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreatePooling2dFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreatePooling2dFloatWorkload)
{
- NeonCreatePooling2dWorkloadTest<NeonPooling2dFloat32Workload, DataType::Float32>();
+ NeonCreatePooling2dWorkloadTest<NeonPooling2dFloatWorkload, DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreatePooling2dUint8Workload)
@@ -298,13 +298,13 @@ static void NeonCreateReshapeWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreateReshapeFloat16Workload)
{
- NeonCreateReshapeWorkloadTest<NeonReshapeFloat32Workload, DataType::Float16>();
+ NeonCreateReshapeWorkloadTest<NeonReshapeFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreateReshapeFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateReshapeFloatWorkload)
{
- NeonCreateReshapeWorkloadTest<NeonReshapeFloat32Workload, DataType::Float32>();
+ NeonCreateReshapeWorkloadTest<NeonReshapeFloatWorkload, DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateReshapeUint8Workload)
@@ -330,20 +330,20 @@ static void NeonCreateSoftmaxWorkloadTest()
#ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC
BOOST_AUTO_TEST_CASE(CreateSoftmaxFloat16Workload)
{
- NeonCreateSoftmaxWorkloadTest<NeonSoftmaxFloat32Workload, DataType::Float16>();
+ NeonCreateSoftmaxWorkloadTest<NeonSoftmaxFloatWorkload, DataType::Float16>();
}
#endif
-BOOST_AUTO_TEST_CASE(CreateSoftmaxFloat32Workload)
+BOOST_AUTO_TEST_CASE(CreateSoftmaxFloatWorkload)
{
- NeonCreateSoftmaxWorkloadTest<NeonSoftmaxFloat32Workload, DataType::Float32>();
+ NeonCreateSoftmaxWorkloadTest<NeonSoftmaxFloatWorkload, DataType::Float32>();
}
BOOST_AUTO_TEST_CASE(CreateSplitterWorkload)
{
Graph graph;
NeonWorkloadFactory factory;
- auto workload = CreateSplitterWorkloadTest<NeonSplitterFloat32Workload, DataType::Float32>(factory, graph);
+ auto workload = CreateSplitterWorkloadTest<NeonSplitterFloatWorkload, DataType::Float32>(factory, graph);
// Checks that outputs are as we expect them (see definition of CreateSplitterWorkloadTest).
SplitterQueueDescriptor queueDescriptor = workload->GetData();
@@ -372,7 +372,7 @@ BOOST_AUTO_TEST_CASE(CreateSplitterMerger)
NeonWorkloadFactory factory;
auto workloads =
- CreateSplitterMergerWorkloadTest<NeonSplitterFloat32Workload, NeonMergerFloat32Workload,
+ CreateSplitterMergerWorkloadTest<NeonSplitterFloatWorkload, NeonMergerFloatWorkload,
DataType::Float32>(factory, graph);
auto wlSplitter = std::move(workloads.first);
@@ -401,14 +401,14 @@ BOOST_AUTO_TEST_CASE(CreateSingleOutputMultipleInputs)
Graph graph;
NeonWorkloadFactory factory;
- std::unique_ptr<NeonSplitterFloat32Workload> wlSplitter;
- std::unique_ptr<NeonActivationFloat32Workload> wlActiv0_0;
- std::unique_ptr<NeonActivationFloat32Workload> wlActiv0_1;
- std::unique_ptr<NeonActivationFloat32Workload> wlActiv1_0;
- std::unique_ptr<NeonActivationFloat32Workload> wlActiv1_1;
-
- CreateSplitterMultipleInputsOneOutputWorkloadTest<NeonSplitterFloat32Workload,
- NeonActivationFloat32Workload, DataType::Float32>(factory, graph, wlSplitter, wlActiv0_0, wlActiv0_1,
+ std::unique_ptr<NeonSplitterFloatWorkload> wlSplitter;
+ std::unique_ptr<NeonActivationFloatWorkload> wlActiv0_0;
+ std::unique_ptr<NeonActivationFloatWorkload> wlActiv0_1;
+ std::unique_ptr<NeonActivationFloatWorkload> wlActiv1_0;
+ std::unique_ptr<NeonActivationFloatWorkload> wlActiv1_1;
+
+ CreateSplitterMultipleInputsOneOutputWorkloadTest<NeonSplitterFloatWorkload,
+ NeonActivationFloatWorkload, DataType::Float32>(factory, graph, wlSplitter, wlActiv0_0, wlActiv0_1,
wlActiv1_0, wlActiv1_1);
armnn::INeonTensorHandle* sOut0 = dynamic_cast<armnn::INeonTensorHandle*>(wlSplitter->GetData().m_Outputs[0]);