From 221f38176b0d4dbc212441779d9bbac3cc0eecfa Mon Sep 17 00:00:00 2001 From: Pablo Tello Date: Wed, 28 Jun 2017 17:27:56 +0100 Subject: COMPMID-421: Fixed FP16 support in Neon GEMM. Fixed GEMM FP16 problem with matrices that are not multiple of 32. Added a new test suite NEON/GEMM/Float16/SmallGEMM. Implemented FP16 function to multiply vector by a matrix. Change-Id: Ie6c692885a48d0206bd6fe748332fa83bc286d67 Reviewed-on: http://mpd-gerrit.cambridge.arm.com/79118 Tested-by: Kaizen Reviewed-by: Moritz Pflanzer --- tests/TensorLibrary.h | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) (limited to 'tests/TensorLibrary.h') diff --git a/tests/TensorLibrary.h b/tests/TensorLibrary.h index 4d7143a206..69b2381171 100644 --- a/tests/TensorLibrary.h +++ b/tests/TensorLibrary.h @@ -505,7 +505,7 @@ void TensorLibrary::fill_tensor_uniform(T &&tensor, std::random_device::result_t fill(tensor, distribution_f16, seed_offset); break; } -#endif +#endif /* ARM_COMPUTE_ENABLE_FP16 */ case DataType::F32: { // It doesn't make sense to check [-inf, inf], so hard code it to a big number -- cgit v1.2.1