From 5057ce9e1866ffa0388543d81af32083b5b1c684 Mon Sep 17 00:00:00 2001 From: Michael Kozlov Date: Wed, 17 Apr 2024 14:34:46 +0100 Subject: Update documentation for 24.04 release Change-Id: Ifec7015ad5712d8b84d65203a5fa21cbefcb04ad Signed-off-by: Michael Kozlov Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/11438 Benchmark: Arm Jenkins Tested-by: Arm Jenkins Reviewed-by: Comments-Addressed: Arm Jenkins --- docs/user_guide/release_version_and_change_log.dox | 9 ++++----- 1 file changed, 4 insertions(+), 5 deletions(-) (limited to 'docs/user_guide/release_version_and_change_log.dox') diff --git a/docs/user_guide/release_version_and_change_log.dox b/docs/user_guide/release_version_and_change_log.dox index 3737dbfc33..b29b81580d 100644 --- a/docs/user_guide/release_version_and_change_log.dox +++ b/docs/user_guide/release_version_and_change_log.dox @@ -43,15 +43,14 @@ If there is more than one release in a month then an extra sequential number is v24.04 Public major release - Add Bfloat16 data type support for @ref NEMatMul. - - Optimize start-up time of @ref NEConvolutionLayer for some input configurations where GeMM is selected as the convolution algorithm - - Optimize @ref NEConvolutionLayer for input tensor size > 1e7 bytes and weight tensor height > 7 - Add support for SoftMax in SME2 for FP32 and FP16. - - Performance optimizations: - - Optimize @ref NESoftmaxLayer for axis != 0 by natively supporting higher axes up to axis 3. - Add support for in place accumulation to CPU GEMM kernels. - Add low-precision Int8 * Int8 -> FP32 CPU GEMM which dequantizes after multiplication - Add is_dynamic flag to QuantizationInfo to signal to operators that it may change after configuration - + - Performance optimizations: + - Optimize start-up time of @ref NEConvolutionLayer for some input configurations where GeMM is selected as the convolution algorithm + - Optimize @ref NEConvolutionLayer for input tensor size > 1e7 bytes and weight tensor height > 7 + - Optimize @ref NESoftmaxLayer for axis != 0 by natively supporting higher axes up to axis 3. v24.02.1 Public patch release - Fix performance regression in fixed-format kernels -- cgit v1.2.1