From fb6aaeb3b02ad3210d57c1d945998f93c11de474 Mon Sep 17 00:00:00 2001 From: Sang-Hoon Park Date: Wed, 27 Nov 2019 15:26:44 +0000 Subject: COMPMID-2773 [NE] add support for QASYMM8_SIGNED to QuantizationLayer Change-Id: Ib692a79228fd85ee600c212d77439ca38d71f332 Signed-off-by: Sang-Hoon Park Reviewed-on: https://review.mlplatform.org/c/2377 Comments-Addressed: Arm Jenkins Tested-by: Arm Jenkins Reviewed-by: Michele Di Giorgio --- arm_compute/core/NEON/NEAsymm.h | 2 -- arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h | 2 +- 2 files changed, 1 insertion(+), 3 deletions(-) (limited to 'arm_compute') diff --git a/arm_compute/core/NEON/NEAsymm.h b/arm_compute/core/NEON/NEAsymm.h index c09a7d9028..e4f4250d16 100644 --- a/arm_compute/core/NEON/NEAsymm.h +++ b/arm_compute/core/NEON/NEAsymm.h @@ -695,7 +695,6 @@ inline uint8x16_t vquantize(const float32x4x4_t &qv, const UniformQuantizationIn * * @return A neon vector holding the quantized values */ - inline int8x16_t vquantize_signed(const float32x4x4_t &qv, const UniformQuantizationInfo &qi) { const float scale = qi.scale; @@ -716,7 +715,6 @@ inline int8x16_t vquantize_signed(const float32x4x4_t &qv, const UniformQuantiza vcvtq_s32_f32(vmlaq_f32(voffset, qv.val[2], vinvscale)), vcvtq_s32_f32(vmlaq_f32(voffset, qv.val[3], vinvscale)), #endif //__aarch64__ - } }; const int8x8_t pa = vqmovn_s16(vcombine_s16(vqmovn_s32(rf.val[0]), vqmovn_s32(rf.val[1]))); diff --git a/arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h b/arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h index 3a50c07ded..1a9b533640 100644 --- a/arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h +++ b/arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h @@ -84,7 +84,7 @@ private: * * @param[in] window Region on which to execute the kernel. */ - template + template void run_quantize_qasymm8(const Window &window); /** Function to apply QASYMM16 quantization on a tensor. * -- cgit v1.2.1