From f948b101d9edb0ea554f175425271b4226ba89de Mon Sep 17 00:00:00 2001 From: Gian Marco Iodice Date: Wed, 14 Aug 2019 07:55:50 +0100 Subject: COMPMID-2582: Disable fuse batch normalization on OpenCL for FP16 Change-Id: I53bf35ff9f6d8c7e3664983ebcc94552a4ebf90d Signed-off-by: Gian Marco Iodice Reviewed-on: https://review.mlplatform.org/c/1732 Tested-by: Arm Jenkins Reviewed-by: Pablo Marquez --- src/graph/mutators/NodeFusionMutator.cpp | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/src/graph/mutators/NodeFusionMutator.cpp b/src/graph/mutators/NodeFusionMutator.cpp index 9e963091e6..f7f3454fad 100644 --- a/src/graph/mutators/NodeFusionMutator.cpp +++ b/src/graph/mutators/NodeFusionMutator.cpp @@ -307,8 +307,10 @@ void NodeFusionMutator::mutate(Graph &g) detail::fuse_layer(g, empty_prec, detail::fuse_node_with_activation, supported_fused_activations); detail::fuse_layer(g, qs8_prec, detail::fuse_node_with_activation, supported_fused_activations); - // TODO (COMPMID-2524): Fuse batch normalization with convolution and depthwise convolution at graph level for NEON - if(target == Target::CL) + // Currently fuse batch normalization brings performance uplift only on OpenCL with FP32 data type + // TODO (COMPMID-2524): Fuse batch normalization with convolution and depthwise convolution at graph level for NEON - FP32 + // TODO (COMPMID-2581): Fuse batch normalization with convolution and depthwise convolution at graph level for OpenCL - FP16 + if(target == Target::CL && (g.nodes()[0].get()->output(0)->desc().data_type == DataType::F32)) { //Depthwise Convolution and Batch Normalization Fusion active only for CL detail::fuse_layer(g, empty_prec, detail::fuse_convolution_with_batch_normalization); -- cgit v1.2.1