From f2cde9b29deee6423ea6fe9a1a9afc9ef61d2663 Mon Sep 17 00:00:00 2001 From: Gian Marco Iodice Date: Thu, 23 Aug 2018 15:29:16 +0100 Subject: COMPMID-1534 - Fix NENormalizationLayer for FP16 Implemented vinvq_f16 with fp32 data type in order to avoid accuracy issue. Change-Id: Ibfffd12e4a941c1388a982fc7bbe3e1832351feb Reviewed-on: https://eu-gerrit-1.euhpc.arm.com/145416 Reviewed-by: Georgios Pinitas Tested-by: Jenkins --- tests/validation/fixtures/NormalizationLayerFixture.h | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) (limited to 'tests/validation/fixtures') diff --git a/tests/validation/fixtures/NormalizationLayerFixture.h b/tests/validation/fixtures/NormalizationLayerFixture.h index f4f9c64944..318b77e1a7 100644 --- a/tests/validation/fixtures/NormalizationLayerFixture.h +++ b/tests/validation/fixtures/NormalizationLayerFixture.h @@ -59,7 +59,8 @@ protected: template void fill(U &&tensor) { - library->fill_tensor_uniform(tensor, 0); + std::uniform_real_distribution<> distribution(-1.0f, 1.0f); + library->fill(tensor, distribution, 0); } TensorType compute_target(const TensorShape &shape, NormalizationLayerInfo info, DataType data_type) -- cgit v1.2.1