aboutsummaryrefslogtreecommitdiff
path: root/src
AgeCommit message (Expand)Author
2022-11-22Remove dynamic fusion prototype with tests and examplesSiCong Li
2022-11-22ONCPUML-1072: Tuned MWS values (for N1, V1) for binary operators used by oneDNNFadi Arafeh
2022-11-18Add num_threads_to_use to OMPScheduler based on workload sizecfRod
2022-11-15Fix regression caused by mws in ActivationLayerMohammed Suhail Munshi
2022-11-15Fixed Arm NN unit test failure caused by quantised multiplication patch.Omar Al Khatib
2022-11-14Optimize Transposed Convolution for CL backend (FP32/16)Gunes Bayir
2022-11-14Optimize T_QUANTIZE8_ASYMMETRIC for Mali™ G52Pablo Marquez Tello
2022-11-10Fix compiler warnings in dynamic fusionSiCong Li
2022-11-09Fix CPU multiplication layer threading overheadViet-Hoa Do
2022-11-08SVE Hard-Swish via Lookup table for quantized inputPablo Marquez Tello
2022-11-07Optimize CPU mul layer on quantized dataOmar Al Khatib
2022-11-04Fix compiler warnings in dynamic fusionSiCong Li
2022-11-03Fix activation block in gemm.clGian Marco Iodice
2022-11-02Partially Revert "Add threshold for floating-point SOFT_RELU activation"Gunes Bayir
2022-11-01Fix fixed-point quantized additionViet-Hoa Do
2022-11-01Updateable weights in depthwise convolutionMilos Puzovic
2022-11-01Add threshold for floating-point SOFT_RELU activationMilos Puzovic
2022-11-01Add check for Batch Matmul in GemmAssemblyDispatchMohammed Suhail Munshi
2022-11-01Rewrite dynamic fusionSiCong Li
2022-11-01Rework direct convolution heuristic on OpenCLGian Marco Iodice
2022-10-27Fix fixed-point quantized additionViet-Hoa Do
2022-10-24Add FP16 tanh based on rational approximationJonathan Deakin
2022-10-20Update reinterpret tensor as 1D for CPU addViet-Hoa Do
2022-10-20Add test in GEMMLowp for batch matmulMohammed Suhail Munshi
2022-10-19Fix FFTConvolutionLayer testViet-Hoa Do
2022-10-12Optimize Neon™ Logistic ActivationMohammed Suhail Munshi
2022-10-12Adding documentation section explaining how BF16 is usedRamy Elgammal
2022-10-10Fix LUT-based activation layerViet-Hoa Do
2022-10-07Workaround CL compiler issue on FP16Viet-Hoa Do
2022-10-07Optimize Neon™ SUB operator by squashing execution windowJakub Sujak
2022-10-06Rework DepthwiseConvolution heuristic on OpenCLGian Marco Iodice
2022-10-06Improve start-up time in gemmlowp reshaped rhs only.Adnan AlSinan
2022-10-04Update GEMM reshaped rhs only heuristicGian Marco Iodice
2022-10-03Force CL kernel compilation with 64 registersViet-Hoa Do
2022-10-03Fix Batch Matmul nightly failureAdnan AlSinan
2022-10-03Optimize CPU add layer on quantized dataViet-Hoa Do
2022-09-28Fix overflow in NEActivationLayer for FP16 typePablo Marquez Tello
2022-09-26Add FP32 Neon™ swish activationJonathan Deakin
2022-09-23CPU GEMM: Fix overreads in SVE merges.David Mansell
2022-09-22Fix unresolved symbol for target armv7a + AndroidPablo Marquez Tello
2022-09-16Fix validation in validate_image2d_support_on_rhsGian Marco Iodice
2022-09-16Fix bug in QASYMM8_SIGNED to F32 cast layerViet-Hoa Do
2022-09-16Optimize Quantized/Integer Bilinear Scale for Neon™Gunes Bayir
2022-09-14Interpreting tensor as 1D for CPU multiplicationViet-Hoa Do
2022-09-14Fix invalid memory access for dynamically fused Cl Elementwise kernelsSiCong Li
2022-09-14Adding GELU activationMurray Kornelsen
2022-09-14INT8 Quantized MeanStdDevNorm (LayerNorm)Murray Kornelsen
2022-09-12Add test for NEGEMM to test a batched matrix multiplication with variable inp...Adnan AlSinan
2022-09-09Rework heuristic in ClConv2dGian Marco Iodice
2022-09-09Optimize FP32/16 Bilinear Scale Kernel for Neon™Gunes Bayir