From 38d93bdda73f9b1024c6b4b834b382f7f25aae19 Mon Sep 17 00:00:00 2001 From: Vidhya Sudhan Loganathan Date: Tue, 20 Nov 2018 15:38:13 +0000 Subject: COMPMID-1801 : (Nightly) CLWinogradConvolutionLayer FP16 mismatches FP mixed precision support added to GEMM kernel used for fp16 winograd conv on Midgard GPUs Change-Id: I1619beb025fc484a1ac9d3e528d785edabbc7ee6 --- tests/benchmark/fixtures/WinogradConvolutionLayerFixture.h | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) (limited to 'tests/benchmark') diff --git a/tests/benchmark/fixtures/WinogradConvolutionLayerFixture.h b/tests/benchmark/fixtures/WinogradConvolutionLayerFixture.h index 5f44517817..86ea9c4dc1 100644 --- a/tests/benchmark/fixtures/WinogradConvolutionLayerFixture.h +++ b/tests/benchmark/fixtures/WinogradConvolutionLayerFixture.h @@ -60,7 +60,11 @@ public: dst = create_tensor(dst_shape, data_type, 1); // Create and configure function - conv_layer.configure(&src, &weights, &biases, &dst, info, act_info); + if(data_type == DataType::F16){ + conv_layer.configure(&src, &weights, &biases, &dst, info, act_info, true); + }else{ + conv_layer.configure(&src, &weights, &biases, &dst, info, act_info); + } // Allocate tensors src.allocator()->allocate(); -- cgit v1.2.1