kpu / intgemm
int8_t and int16_t matrix multiply based on https://arxiv.org/abs/1705.01991
☆68Updated last year
Alternatives and similar repositories for intgemm:
Users that are interested in intgemm are comparing it to the libraries listed below
- Fast matrix multiplication for few-bit integer matrices on CPUs.☆27Updated 6 years ago
- C99/C++ header-only library for division via fixed-point multiplication by inverse☆50Updated 11 months ago
- how to design cpu gemm on x86 with avx256, that can beat openblas.☆68Updated 5 years ago
- ☆309Updated 3 months ago
- Fast stand-alone C++ decoder for RNN-based NMT models☆25Updated 4 years ago
- Fast Neural Machine Translation in C++ - development repository☆268Updated 5 months ago
- A library of GPU kernels for sparse matrix operations.☆260Updated 4 years ago
- CUDA templates for tile-sparse matrix multiplication based on CUTLASS.☆50Updated 7 years ago
- A GPU language model, based on btree backed tries.☆30Updated 7 years ago
- Codebase associated with the PyTorch compiler tutorial☆46Updated 5 years ago
- Benchmarks to capture important workloads.☆30Updated last month
- MatMul Performance Benchmarks for a Single CPU Core comparing both hand engineered and codegen kernels.☆129Updated last year
- Customized matrix multiplication kernels☆53Updated 3 years ago
- THIS REPOSITORY HAS MOVED TO github.com/nvidia/cub, WHICH IS AUTOMATICALLY MIRRORED HERE.☆84Updated last year
- Stretching GPU performance for GEMMs and tensor contractions.☆233Updated last week
- Training neural networks in TensorFlow 2.0 with 5x less memory☆130Updated 3 years ago
- Clover: Quantized 4-bit Linear Algebra Library☆112Updated 6 years ago
- Conversion to/from half-precision floating point formats☆345Updated 7 months ago
- Kernel Fusion and Runtime Compilation Based on NNVM☆70Updated 8 years ago
- Library for fast image convolution in neural networks on Intel Architecture☆29Updated 7 years ago
- Applications using the GTN library and code to reproduce experiments in "Differentiable Weighted Finite-State Transducers"☆83Updated 2 years ago
- Personal collection of references for high performance mixed precision training.☆41Updated 5 years ago
- Fast integer division with divisor not known at compile time. To be used primarily in CUDA kernels.☆70Updated 9 years ago
- Intel® Optimization for Chainer*, a Chainer module providing numpy like API and DNN acceleration using MKL-DNN.☆168Updated last week
- nGraph™ Backend for ONNX☆42Updated 2 years ago
- ☆69Updated 2 years ago
- Test winograd convolution written in TVM for CUDA and AMDGPU☆41Updated 6 years ago
- Benchmark code for the "Online normalizer calculation for softmax" paper☆85Updated 6 years ago
- DLPack for Tensorflow☆36Updated 4 years ago
- ☆49Updated last year