neuralmagic / sparsify
ML model optimization product to accelerate inference.
☆327Updated this week
Alternatives and similar repositories for sparsify:
Users that are interested in sparsify are comparing it to the libraries listed below
- Neural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes☆383Updated this week
- Top-level directory for documentation and general content☆121Updated 5 months ago
- Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models☆2,131Updated this week
- Sparsity-aware deep learning inference runtime for CPUs☆3,143Updated this week
- Accelerate PyTorch models with ONNX Runtime☆359Updated 2 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆263Updated 6 months ago
- Library for 8-bit optimizers and quantization routines.☆716Updated 2 years ago
- DiffQ performs differentiable quantization using pseudo quantization noise. It can automatically tune the number of bits used per weight …☆235Updated 2 years ago
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,042Updated last year
- An open-source efficient deep learning framework/compiler, written in python.☆698Updated 2 months ago
- Recipes are a standard, well supported set of blueprints for machine learning engineers to rapidly train models using the latest research…☆312Updated this week
- Easily benchmark PyTorch model FLOPs, latency, throughput, allocated gpu memory and energy consumption☆102Updated last year
- Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackab…☆1,566Updated last year
- Curated list of awesome material on optimization techniques to make artificial intelligence faster and more efficient 🚀☆113Updated last year
- Implementation of a Transformer, but completely in Triton☆264Updated 3 years ago
- A research library for pytorch-based neural network pruning, compression, and more.☆161Updated 2 years ago
- A library that contains a rich collection of performant PyTorch model metrics, a simple interface to create new metrics, a toolkit to fac…☆228Updated 3 months ago
- Accelerate your Neural Architecture Search (NAS) through fast, reproducible and modular research.☆476Updated 6 months ago
- Productionize machine learning predictions, with ONNX or without☆65Updated last year
- Fast Block Sparse Matrices for Pytorch☆545Updated 4 years ago
- Prune a model while finetuning or training.☆402Updated 2 years ago
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆462Updated this week
- A library to inspect and extract intermediate layers of PyTorch models.☆473Updated 2 years ago
- The merlin dataloader lets you rapidly load tabular data for training deep leaning models with TensorFlow, PyTorch or JAX☆418Updated last year
- Lite Inference Toolkit (LIT) for PyTorch☆161Updated 3 years ago
- FasterAI: Prune and Distill your models with FastAI and PyTorch☆248Updated last month
- An open-source AutoML Library based on PyTorch☆307Updated last month
- Fast sparse deep learning on CPUs☆53Updated 2 years ago
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆348Updated this week
- A repository for log-time feedforward networks☆222Updated last year