From 659abc022b42ad8d11f58b3f3b7b4175f3cf1038 Mon Sep 17 00:00:00 2001 From: Pablo Tello Date: Thu, 22 Jun 2017 16:00:16 +0100 Subject: COMPMID-421: Added FP16 support in Convolutional layer (Neon) The test suite for FP16 is conditionally compiled in when the target platform is arch=arm64-8.2-a Change-Id: I1686157e83809a00a91058bff80dbecf692fb356 Reviewed-on: http://mpd-gerrit.cambridge.arm.com/78740 Tested-by: Kaizen Reviewed-by: Anthony Barbier --- src/core/NEON/kernels/NEWeightsReshapeKernel.cpp | 11 ++++++++--- 1 file changed, 8 insertions(+), 3 deletions(-) (limited to 'src/core/NEON/kernels/NEWeightsReshapeKernel.cpp') diff --git a/src/core/NEON/kernels/NEWeightsReshapeKernel.cpp b/src/core/NEON/kernels/NEWeightsReshapeKernel.cpp index f31cde719a..4f52bf6279 100644 --- a/src/core/NEON/kernels/NEWeightsReshapeKernel.cpp +++ b/src/core/NEON/kernels/NEWeightsReshapeKernel.cpp @@ -95,7 +95,7 @@ NEWeightsReshapeKernel::NEWeightsReshapeKernel() void NEWeightsReshapeKernel::configure(const ITensor *input, const ITensor *bias, ITensor *output) { - ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(input, 1, DataType::F32, DataType::QS8); + ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(input, 1, DataType::F32, DataType::F16, DataType::QS8); ARM_COMPUTE_ERROR_ON_NULLPTR(output); ARM_COMPUTE_ERROR_ON(input->info()->dimension(0) != input->info()->dimension(1)); @@ -113,7 +113,7 @@ void NEWeightsReshapeKernel::configure(const ITensor *input, const ITensor *bias set_fixed_point_position_if_zero(*output->info(), fixed_point_position); ARM_COMPUTE_ERROR_ON_MISMATCHING_DIMENSIONS(output->info()->tensor_shape(), output_shape); - ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(output, 1, DataType::F32, DataType::QS8); + ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(output, 1, DataType::F16, DataType::F32, DataType::QS8); ARM_COMPUTE_ERROR_ON_MISMATCHING_DATA_TYPES(input, output); ARM_COMPUTE_ERROR_ON_MISMATCHING_FIXED_POINT(input, output); @@ -127,7 +127,7 @@ void NEWeightsReshapeKernel::configure(const ITensor *input, const ITensor *bias set_shape_if_empty(*bias->info(), bias_shape); ARM_COMPUTE_ERROR_ON_MISMATCHING_DIMENSIONS(bias->info()->tensor_shape(), bias_shape); - ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(bias, 1, DataType::F32, DataType::QS8); + ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(bias, 1, DataType::F16, DataType::F32, DataType::QS8); ARM_COMPUTE_ERROR_ON_MISMATCHING_DATA_TYPES(input, bias); ARM_COMPUTE_ERROR_ON_MISMATCHING_FIXED_POINT(input, output); } @@ -143,6 +143,11 @@ void NEWeightsReshapeKernel::configure(const ITensor *input, const ITensor *bias _func = &weights_reshape; break; } + case DataType::F16: + { + _func = &weights_reshape; + break; + } case DataType::QS8: { _func = &weights_reshape; -- cgit v1.2.1