From a25d16c86f0d870408bc8b941aa755093417b0f0 Mon Sep 17 00:00:00 2001 From: Vidhya Sudhan Loganathan Date: Fri, 16 Nov 2018 11:33:12 +0000 Subject: COMPMID-1266 : Add support for FP16 in CLWinogradConvolutionLayer: 5x5 kernels Introduced F32 accumulation for F16 winograd gemm and output transform WinogradConvolution will be available for F16 only if fast math flag is enabled Change-Id: I215593c205236a0f9669218437bb40b184ec6a4f --- src/core/CL/CLKernelLibrary.cpp | 1 + 1 file changed, 1 insertion(+) (limited to 'src/core/CL/CLKernelLibrary.cpp') diff --git a/src/core/CL/CLKernelLibrary.cpp b/src/core/CL/CLKernelLibrary.cpp index fde9608949..955844da3e 100644 --- a/src/core/CL/CLKernelLibrary.cpp +++ b/src/core/CL/CLKernelLibrary.cpp @@ -256,6 +256,7 @@ const std::map CLKernelLibrary::_kernel_program_map = { "gemm_mm_interleaved_transposed_f32_bifrost", "gemm.cl" }, { "gemm_mm_floating_point", "gemm.cl" }, { "gemm_mm_floating_point_f16_bifrost", "gemm.cl" }, + { "gemm_mm_floating_point_f16_bifrost_acc32", "gemm.cl" }, { "gemm_mm_floating_point_f32_bifrost", "gemm.cl" }, { "gemm_mm_floating_point_f32_bifrost_1000", "gemm.cl" }, { "gemm_lc_vm_f32", "gemm.cl" }, -- cgit v1.2.1