diff options
author | Georgios Pinitas <georgios.pinitas@arm.com> | 2018-07-12 10:49:53 +0100 |
---|---|---|
committer | Anthony Barbier <anthony.barbier@arm.com> | 2018-11-02 16:54:12 +0000 |
commit | b03f7c5c780fe2df23eb8c5c1b4b1d65bd7f0339 (patch) | |
tree | 91130cb09309fbc526de48b2bc538fc71f80f4e8 /src/runtime | |
parent | 12be7ab4876f77fecfab903df70791623219b3da (diff) | |
download | ComputeLibrary-b03f7c5c780fe2df23eb8c5c1b4b1d65bd7f0339.tar.gz |
COMPMID-1188 : Rename TNOX to Mali-G76
Change-Id: I136f7aa4bca268abd4fbe4f6ce4bcc2708ec3671
Reviewed-on: https://eu-gerrit-1.euhpc.arm.com/139689
Tested-by: Jenkins <bsgcomp@arm.com>
Reviewed-by: Anthony Barbier <anthony.barbier@arm.com>
Diffstat (limited to 'src/runtime')
-rw-r--r-- | src/runtime/CL/functions/CLGEMM.cpp | 2 | ||||
-rw-r--r-- | src/runtime/CL/functions/CLGEMMLowpMatrixMultiplyCore.cpp | 2 | ||||
-rw-r--r-- | src/runtime/CL/tuners/BifrostTuner.cpp | 12 |
3 files changed, 8 insertions, 8 deletions
diff --git a/src/runtime/CL/functions/CLGEMM.cpp b/src/runtime/CL/functions/CLGEMM.cpp index bb76872700..1f4df4f1a9 100644 --- a/src/runtime/CL/functions/CLGEMM.cpp +++ b/src/runtime/CL/functions/CLGEMM.cpp @@ -44,7 +44,7 @@ inline bool is_interleaved_transposed(int m, int n, int k, DataType data_type, b { bool flag = true; - if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72)) + if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G76)) { // COMPMID-852 if(k > 256 && m > 4 && is_data_type_float(data_type) && reshape_b_only_on_first_run) diff --git a/src/runtime/CL/functions/CLGEMMLowpMatrixMultiplyCore.cpp b/src/runtime/CL/functions/CLGEMMLowpMatrixMultiplyCore.cpp index 94dc0e071c..842ee73397 100644 --- a/src/runtime/CL/functions/CLGEMMLowpMatrixMultiplyCore.cpp +++ b/src/runtime/CL/functions/CLGEMMLowpMatrixMultiplyCore.cpp @@ -41,7 +41,7 @@ inline bool is_interleaved_transposed(int m, int n, int k, bool reshape_b_only_o { bool flag = true; - if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::TNOX)) + if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::G76)) { // COMPMID-852 if(k > 256 && m > 4 && reshape_b_only_on_first_run) diff --git a/src/runtime/CL/tuners/BifrostTuner.cpp b/src/runtime/CL/tuners/BifrostTuner.cpp index edd074ba08..fa67710cc8 100644 --- a/src/runtime/CL/tuners/BifrostTuner.cpp +++ b/src/runtime/CL/tuners/BifrostTuner.cpp @@ -132,7 +132,7 @@ void tune_col2im_kernel(CLCol2ImKernel &k) // Configure the local work size for Bifrost with a value obtained // via exhaustive autotuning over 30 representative tensor shapes. - if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::TNOX)) + if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::G76)) { if((k._convolved_dims.first == 7) || (k._convolved_dims.first == 14)) { @@ -153,7 +153,7 @@ void tune_im2col_kernel(CLIm2ColKernel &k) const GPUTarget gpu_target = k.get_target(); // Local work size optimized for the 11x11 AlexNet convolution on Bifrost. - if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::TNOX) && k._kernel_dims.width == 11) + if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::G76) && k._kernel_dims.width == 11) { const bool is_square_kernel = (k._kernel_dims.width == k._kernel_dims.height); if(!is_square_kernel && k._kernel_dims.width > 1 && !k._conv_info.has_padding()) @@ -171,7 +171,7 @@ void tune_depthwise_im2col_kernel(CLDepthwiseIm2ColKernel &k) // Configure the local work size for Bifrost with a value obtained // via exhaustive autotuning for the MobileNets tensor shapes. - if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::TNOX)) + if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::G76)) { lws_hint = cl::NDRange(1, 2, 1); } @@ -186,7 +186,7 @@ void tune_gemv_kernel(CLGEMMMatrixVectorMultiplyKernel &k) // Configure the local work size for Bifrost with a value obtained // via exhaustive autotuning for the MobileNets tensor shapes. - if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::TNOX)) + if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::G76)) { lws_hint = cl::NDRange(1, 1, 1); } @@ -207,7 +207,7 @@ void tune_gemm_kernel(CLGEMMMatrixMultiplyKernel &k) case GPUTarget::G51: case GPUTarget::G51BIG: case GPUTarget::G51LIT: - case GPUTarget::TNOX: + case GPUTarget::G76: if(k._input1->info()->dimension(1) == 24) { // LWS optimized for the 11x11 AlexNet convolution on Bifrost. @@ -240,7 +240,7 @@ void tune_pooling_kernel(CLPoolingLayerKernel &k) // invalid (e.g. exceeds the maximum workgroup size that the kernel can be launched with). if(k._input->info()->data_layout() == DataLayout::NCHW) { - if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::TNOX)) + if(gpu_target_is_in(gpu_target, GPUTarget::G71, GPUTarget::G72, GPUTarget::G51, GPUTarget::G51BIG, GPUTarget::G51LIT, GPUTarget::G76)) { cl::NDRange gws = ICLKernel::gws_from_window(k.window()); lws_hint = cl::NDRange(gws[0], gws[1], 1); |