From ee209d8d3cac5fbdc724c2f23628d597974e96e8 Mon Sep 17 00:00:00 2001 From: Dominic Symes Date: Thu, 1 Apr 2021 16:14:08 +0100 Subject: Remove RELUN operation The operation can be implemented directly with CLAMP for int8, int16 or MIN,MAX for int32. Change-Id: I872ec5ee4ee30bb69353f39be12a65307c9b8d0a Signed-off-by: Dominic Symes --- chapters/activation_funcs.adoc | 34 ---------------------------------- 1 file changed, 34 deletions(-) diff --git a/chapters/activation_funcs.adoc b/chapters/activation_funcs.adoc index 5c8e36f..a27d1bf 100644 --- a/chapters/activation_funcs.adoc +++ b/chapters/activation_funcs.adoc @@ -44,40 +44,6 @@ for_each(index in shape) { |MI, MT|floating-point|float_t|float_t |=== -==== RELUN - -ReLU with a scalar maximum value. - -*Arguments:* - -|=== -|Argument|Type|Name|Shape|Description - -|Input|in_t*|Input|shape|Input tensor -|Attribute|in_t|max_val|-|maximum clip value -|Output|in_t*|Output|shape|Output tensor of same type and shape as input -|=== - -*Operation Function:* - -[source,c++] ----- -for_each(index in shape) { - in_t value = tensor_read(input, shape, index); - value = apply_clip(value, 0, max_val); - tensor_write(output, shape, index, value); -} ----- - -*Supported Data Types:* - -|=== -|Profile|Mode|in_t - -|Any|signed 32|int32_t -|MI, MT|floating-point|float_t -|=== - ==== SIGMOID Sigmoid function: output = 1 / (1 + exp(-input)) -- cgit v1.2.1