ARM-software / ComputeLibraryLinks
The Compute Library is a set of computer vision and machine learning functions optimised for both Arm CPUs and GPUs using SIMD technologies.
☆3,080Updated this week
Alternatives and similar repositories for ComputeLibrary
Users that are interested in ComputeLibrary are comparing it to the libraries listed below
Sorting:
- Arm NN ML Software.☆1,290Updated 2 weeks ago
- An open optimized software library project for the ARM® Architecture☆1,516Updated 3 years ago
- Low-precision matrix multiplication☆1,819Updated last year
- C++ image processing and machine learning library with using of SIMD: SSE, AVX, AVX-512, AMX for x86/x64, NEON for ARM.☆2,218Updated last week
- Quantized Neural Network PACKage - mobile-optimized implementation of quantized neural network operators☆1,547Updated 6 years ago
- oneAPI Deep Neural Network Library (oneDNN)☆3,933Updated this week
- Makes ARM NEON documentation accessible (with examples)☆404Updated last year
- High-efficiency floating-point neural network inference operators for mobile, server, and Web☆2,187Updated this week
- Acceleration package for neural networks on multi-core CPUs☆1,702Updated last year
- ☆1,966Updated 2 years ago
- Tuned OpenCL BLAS☆1,162Updated last week
- The platform independent header allowing to compile any C/C++ code containing ARM NEON intrinsic functions for x86 target systems using S…☆478Updated last month
- Khronos OpenVX Tutorial Material☆246Updated 4 years ago
- Compiler for Neural Network hardware accelerators☆3,323Updated last year
- Arm Machine Learning tutorials and examples☆477Updated this week
- [ARCHIVED] Cooperative primitives for CUDA C++. See https://github.com/NVIDIA/cccl☆1,808Updated 2 years ago
- pocl - Portable Computing Language☆1,039Updated last week
- ☆156Updated 9 months ago
- FeatherCNN is a high performance inference engine for convolutional neural networks.☆1,223Updated 6 years ago
- a language for fast, portable data-parallel computation☆6,451Updated this week
- Bolt is a deep learning library with high performance and heterogeneous flexibility.☆954Updated 8 months ago
- Tengine is a lite, high performance, modular inference engine for embedded device☆4,501Updated 9 months ago
- Khronos OpenCL-Headers☆743Updated last month
- A retargetable MLIR-based machine learning compiler and runtime toolkit.☆3,488Updated this week
- High performance Cross-platform Inference-engine, you could run Anakin on x86-cpu,arm, nv-gpu, amd-gpu,bitmain and cambricon devices.☆535Updated 3 years ago
- Assembler for NVIDIA Maxwell architecture☆1,054Updated 2 years ago
- This fork of BVLC/Caffe is dedicated to improving performance of this deep learning framework when running on CPU, in particular Intel® X…☆851Updated 3 years ago
- Compute Library for Deep Neural Networks (clDNN)☆575Updated 2 years ago
- Library for specialized dense and sparse matrix operations, and deep learning primitives.☆923Updated last week
- Optimized implementations of various library functions for ARM architecture processors☆669Updated last week