intel / light-model-transformer
☆71Updated 2 months ago
Alternatives and similar repositories for light-model-transformer:
Users that are interested in light-model-transformer are comparing it to the libraries listed below
- Library for fast image convolution in neural networks on Intel Architecture☆29Updated 7 years ago
- Accelerating DNN Convolutional Layers with Micro-batches☆64Updated 4 years ago
- High Efficiency Convolution Kernel for Maxwell GPU Architecture☆134Updated 7 years ago
- DNN Inference with CPU, C++, ONNX support: Instant☆56Updated 6 years ago
- A prototype implementation of AllReduce collective communication routine.☆19Updated 6 years ago
- Symbolic Expression and Statement Module for new DSLs☆205Updated 4 years ago
- Intel® Optimization for Chainer*, a Chainer module providing numpy like API and DNN acceleration using MKL-DNN.☆166Updated 2 weeks ago
- A simple memory manager for CUDA designed to help Deep Learning frameworks manage memory☆296Updated 6 years ago
- Optimized half precision gemm assembly kernels (deprecated due to ROCm)☆47Updated 7 years ago
- Kernel Fusion and Runtime Compilation Based on NNVM☆70Updated 8 years ago
- Chainer x TensorRT☆34Updated 5 years ago
- flexible-gemm conv of deepcore☆17Updated 5 years ago
- Documentation for StreamExecutor open source proposal☆83Updated 8 years ago
- Code for testing the native float16 matrix multiplication performance on Tesla P100 and V100 GPU based on cublasHgemm☆34Updated 5 years ago
- Test winograd convolution written in TVM for CUDA and AMDGPU☆40Updated 6 years ago
- Python bindings for NVTX☆66Updated last year
- Greentea LibDNN - a universal convolution implementation supporting CUDA and OpenCL☆135Updated 7 years ago
- Subpart source code of of deepcore v0.7☆27Updated 4 years ago
- Quantize weights and activations in Recurrent Neural Networks.☆94Updated 6 years ago
- TensorFlow and TVM integration☆37Updated 4 years ago
- Menoh: fast DNN inference library with multiple programming language support☆280Updated 4 years ago
- Conversion to/from half-precision floating point formats☆341Updated 5 months ago
- int8_t and int16_t matrix multiply based on https://arxiv.org/abs/1705.01991☆67Updated last year
- THIS REPOSITORY HAS MOVED TO github.com/nvidia/cub, WHICH IS AUTOMATICALLY MIRRORED HERE.☆83Updated 11 months ago
- Fast matrix multiplication for few-bit integer matrices on CPUs.☆27Updated 5 years ago
- Tools and extensions for CUDA profiling☆63Updated 5 years ago
- Efficient Top-K implementation on the GPU☆150Updated 5 years ago
- oneCCL Bindings for Pytorch*☆87Updated 3 weeks ago
- Efficient forward propagation for BCNNs☆50Updated 7 years ago
- tutorial to optimize GEMM performance on android☆51Updated 8 years ago