From 77bbe2e08b0376edfd3f504950be7f4b5720eeb0 Mon Sep 17 00:00:00 2001 From: Viet-Hoa Do Date: Wed, 6 Dec 2023 11:01:15 +0000 Subject: Add SME2 implementation of softmax for FP32 Signed-off-by: Viet-Hoa Do Change-Id: I8a63610cfb9ccff89dec6045d023439fc19b027a Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/11357 Tested-by: Arm Jenkins Reviewed-by: Gunes Bayir Comments-Addressed: Arm Jenkins Benchmark: Arm Jenkins --- docs/user_guide/release_version_and_change_log.dox | 1 + 1 file changed, 1 insertion(+) (limited to 'docs/user_guide') diff --git a/docs/user_guide/release_version_and_change_log.dox b/docs/user_guide/release_version_and_change_log.dox index 31b756070d..aa27c2b44c 100644 --- a/docs/user_guide/release_version_and_change_log.dox +++ b/docs/user_guide/release_version_and_change_log.dox @@ -45,6 +45,7 @@ v24.04 Public major release - Add Bfloat16 data type support for @ref NEMatMul. - Optimize start-up time of @ref NEConvolutionLayer for some input configurations where GeMM is selected as the convolution algorithm - Optimize @ref NEConvolutionLayer for input tensor size > 1e7 bytes and weight tensor height > 7 + - Add support for SoftMax in SME2 for FP32. - Performance optimizations: - Optimize @ref NESoftmaxLayer for axis != 0 by natively supporting higher axes up to axis 3. -- cgit v1.2.1