aboutsummaryrefslogtreecommitdiff
path: root/src/cpu
AgeCommit message (Expand)Author
2023-05-03Fix CPU MatMul broadcast detectionViet-Hoa Do
2023-05-02Fix fully connected and matmul mismatchesViet-Hoa Do
2023-04-26Integrate multi-threaded pretranspose_B_arraySiCong Li
2023-04-19NETranspose 8x8 kernel for 32-bit elementsEthan Doe
2023-04-19Add quantized support for CPU MatMulViet-Hoa Do
2023-04-17Fix unhandled case in ElementwiseUnaryRamy Elgammal
2023-04-14Fix dynamic weights for CPU connected layerViet-Hoa Do
2023-04-13Implement MatMul Function and Operator with Floating Point support for CPUMohammed Suhail Munshi
2023-04-11Enable quantized data types for CpuElementwiseUnary on Armv7aRamy Elgammal
2023-03-29Fix GCC13 compiler errorsPablo Marquez Tello
2023-03-29Add quantized support for unary elementwise in CPUViet-Hoa Do
2023-03-23Round to nearest with ties to away from zero in ReluPablo Marquez Tello
2023-03-21Add dynamic weights for CPU fully connected layerViet-Hoa Do
2023-03-13[ONCPUML-1174] Allow src/weights mismatch for fixed formatJonathan Deakin
2023-03-03NEGEMMLowpMatrixMultiplyCore should be configured for optimized int8 kernel.Ethan Doe
2023-03-01Add support for kernel indices in MaxpoolAdnan AlSinan
2023-02-28Add an option to use lowest for max-poolingAdnan AlSinan
2023-02-08Update CPU kernels to remove x19 and w19Michael Tyler
2023-02-08Add support for dilation > 1 in assembly DepthwiseConvolutionPablo Marquez Tello
2023-02-03Fix armv7a failing GEMMConvolutionLayer testsMohammed Suhail Munshi
2023-02-01Fix GEMMLowp/Batched MatMul mismatches on CPUMohammed Suhail Munshi
2023-02-01Add new operator AddMulAdd for Neon™ backend for Float/Quantized typesGunes Bayir
2023-02-01Remove fixed format strides hackJonathan Deakin
2023-01-18Add broadcast batched matmul validation casesSiCong Li
2023-01-18Revert "Update CPU kernels to remove x19"Michael Tyler
2023-01-16Update CPU kernels to remove x19Michael Tyler
2023-01-11Deprecated BF16 support in DepthConvertPablo Marquez Tello
2022-12-29Use CPU quantized addition kernel for quantized subtractionOmar Al Khatib
2022-12-21Fixed various mismatches in CpuCastKernelPablo Marquez Tello
2022-11-30Fix build error for unused variables in data type specific buildsGunes Bayir
2022-11-23ONCPUML-1072: Remove double definition of get_mws for Mul kernelfadara01
2022-11-22ONCPUML-1072: Tuned MWS values (for N1, V1) for binary operators used by oneDNNFadi Arafeh
2022-11-15Fix regression caused by mws in ActivationLayerMohammed Suhail Munshi
2022-11-15Fixed Arm NN unit test failure caused by quantised multiplication patch.Omar Al Khatib
2022-11-09Fix CPU multiplication layer threading overheadViet-Hoa Do
2022-11-08SVE Hard-Swish via Lookup table for quantized inputPablo Marquez Tello
2022-11-07Optimize CPU mul layer on quantized dataOmar Al Khatib
2022-11-01Fix fixed-point quantized additionViet-Hoa Do
2022-11-01Updateable weights in depthwise convolutionMilos Puzovic
2022-11-01Add threshold for floating-point SOFT_RELU activationMilos Puzovic
2022-11-01Add check for Batch Matmul in GemmAssemblyDispatchMohammed Suhail Munshi
2022-10-27Fix fixed-point quantized additionViet-Hoa Do
2022-10-20Update reinterpret tensor as 1D for CPU addViet-Hoa Do
2022-10-20Add test in GEMMLowp for batch matmulMohammed Suhail Munshi
2022-10-19Fix FFTConvolutionLayer testViet-Hoa Do
2022-10-12Optimize Neon™ Logistic ActivationMohammed Suhail Munshi
2022-10-12Adding documentation section explaining how BF16 is usedRamy Elgammal
2022-10-10Fix LUT-based activation layerViet-Hoa Do
2022-10-07Optimize Neon™ SUB operator by squashing execution windowJakub Sujak
2022-10-03Fix Batch Matmul nightly failureAdnan AlSinan