From fb0fdcdaec57e6f8e1b96f924411921cc0ba6d94 Mon Sep 17 00:00:00 2001 From: Georgios Pinitas Date: Thu, 22 Aug 2019 17:10:04 +0100 Subject: COMPMID-2460: Add ELU activation on NEON,CL,GLES Change-Id: I641b12fcfc5dc4a18d2d489c38673b937125dfb9 Signed-off-by: Georgios Pinitas Reviewed-on: https://review.mlplatform.org/c/1796 Tested-by: Arm Jenkins Reviewed-by: Michele Di Giorgio Comments-Addressed: Arm Jenkins --- src/core/GLES_COMPUTE/cs_shaders/activation_layer.cs | 7 ++++++- src/core/GLES_COMPUTE/cs_shaders/activation_layer_helpers_cs.h | 5 +++++ 2 files changed, 11 insertions(+), 1 deletion(-) (limited to 'src/core/GLES_COMPUTE') diff --git a/src/core/GLES_COMPUTE/cs_shaders/activation_layer.cs b/src/core/GLES_COMPUTE/cs_shaders/activation_layer.cs index dd97c1501b..e5411de985 100644 --- a/src/core/GLES_COMPUTE/cs_shaders/activation_layer.cs +++ b/src/core/GLES_COMPUTE/cs_shaders/activation_layer.cs @@ -69,6 +69,8 @@ void main(void) data_out = lrelu_op(data); #elif defined(SRELU) /*SRELU*/ data_out = srelu_op(data); +#elif defined(ELU) /*ELU*/ + data_out = elu_op(data); #elif defined(ABS) /*ABS*/ data_out = abs_op(data); #elif defined(SQUARE) /*SQUARE*/ @@ -121,6 +123,9 @@ void main(void) #elif defined(SRELU) /*SRELU*/ data_out.x = srelu_op(a); data_out.y = srelu_op(b); +#elif defined(ELU) /*ELU*/ + data_out.x = elu_op(a); + data_out.y = elu_op(b); #elif defined(ABS) /*ABS*/ data_out.x = abs_op(a); data_out.y = abs_op(b); @@ -133,7 +138,7 @@ void main(void) #elif defined(LINEAR) /*LINEAR*/ data_out.x = linear_op(a); data_out.y = linear_op(b); -#elif defined(IDENTITY) /*IDENTITY*/ +#elif defined(IDENTITY) /*IDENTITY*/ data_out.x = identity_op(a); data_out.y = identity_op(b); #else /*LOGISTIC*/ diff --git a/src/core/GLES_COMPUTE/cs_shaders/activation_layer_helpers_cs.h b/src/core/GLES_COMPUTE/cs_shaders/activation_layer_helpers_cs.h index e5a89a830f..e0eacf8dc4 100644 --- a/src/core/GLES_COMPUTE/cs_shaders/activation_layer_helpers_cs.h +++ b/src/core/GLES_COMPUTE/cs_shaders/activation_layer_helpers_cs.h @@ -97,6 +97,11 @@ float srelu_op(float x) { return LOG_OP(ADD_OP(CONST_ONE, EXP_OP(x))); } +// ELU Activation +float elu_op(float x) +{ + return (x >= float(0.0)) ? x : MUL_OP(float(A_VAL), SUB_OP(EXP_OP(x), CONST_ONE)); +} // Absolute Activation float abs_op(float x) { -- cgit v1.2.1