From fe879118c8ced2c84915cc3acedb28cc98a9525d Mon Sep 17 00:00:00 2001 From: Georgios Pinitas Date: Mon, 25 Feb 2019 14:33:25 +0000 Subject: COMPMID-1710: Remove Soft Relu activation from large runs. Soft relu results in calculating infinities when running in large shapes, thus disabling for now. Change-Id: I61eb4d3198c69d2762001299228836e445802509 Signed-off-by: Georgios Pinitas Reviewed-on: https://review.mlplatform.org/c/784 Tested-by: Arm Jenkins Reviewed-by: Michalis Spyrou --- tests/validation/CL/Winograd.cpp | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/tests/validation/CL/Winograd.cpp b/tests/validation/CL/Winograd.cpp index 3f38cfd24c..e744473b34 100644 --- a/tests/validation/CL/Winograd.cpp +++ b/tests/validation/CL/Winograd.cpp @@ -147,14 +147,14 @@ const auto ActivationFunctionsDataset = framework::dataset::make("ActivationInfo ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::RELU), ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::BOUNDED_RELU), ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LU_BOUNDED_RELU), - ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LEAKY_RELU), - ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::SOFT_RELU) + ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LEAKY_RELU) }); const auto ActivationFunctionsSmallDataset = framework::dataset::make("ActivationInfo", { ActivationLayerInfo(), ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LU_BOUNDED_RELU), - ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LEAKY_RELU) + ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::LEAKY_RELU), + ActivationLayerInfo(ActivationLayerInfo::ActivationFunction::SOFT_RELU) }); } // namespace -- cgit v1.2.1