neuralmagic / sparsezoo
Neural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes
☆371Updated 3 months ago
Related projects ⓘ
Alternatives and complementary repositories for sparsezoo
- ML model optimization product to accelerate inference.☆320Updated 6 months ago
- Top-level directory for documentation and general content☆120Updated 4 months ago
- Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models☆2,064Updated 3 months ago
- Sparsity-aware deep learning inference runtime for CPUs☆3,011Updated 3 months ago
- An open-source efficient deep learning framework/compiler, written in python.☆649Updated this week
- DiffQ performs differentiable quantization using pseudo quantization noise. It can automatically tune the number of bits used per weight …☆234Updated last year
- Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. Th…☆324Updated this week
- ☆262Updated this week
- Highly optimized inference engine for Binarized Neural Networks☆243Updated last week
- An Open-Source Library for Training Binarized Neural Networks☆707Updated 2 months ago
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,010Updated 6 months ago
- Prune a model while finetuning or training.☆394Updated 2 years ago
- Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackab…☆1,532Updated 8 months ago
- A research library for pytorch-based neural network pruning, compression, and more.☆160Updated last year
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆280Updated this week
- A library for researching neural networks compression and acceleration methods.☆136Updated 2 months ago
- Recipes are a standard, well supported set of blueprints for machine learning engineers to rapidly train models using the latest research…☆293Updated this week
- Library for 8-bit optimizers and quantization routines.☆714Updated 2 years ago
- Reference implementations of popular Binarized Neural Networks☆104Updated last week
- Accelerate PyTorch models with ONNX Runtime☆355Updated 2 months ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆939Updated this week
- A repository for log-time feedforward networks☆216Updated 7 months ago
- FasterAI: Prune and Distill your models with FastAI and PyTorch☆243Updated last week
- OTOv1-v3, NeurIPS, ICLR, TMLR, DNN Training, Compression, Structured Pruning, Erasing Operators, CNN, Diffusion, LLM☆292Updated last month
- PyTorch library to facilitate development and standardized evaluation of neural network pruning methods.☆423Updated last year
- Pipeline Parallelism for PyTorch☆725Updated 2 months ago
- This repository contains the experimental PyTorch native float8 training UX☆211Updated 3 months ago
- [ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models☆1,242Updated 3 months ago
- Implementation of a Transformer, but completely in Triton☆248Updated 2 years ago
- TensorRT Model Optimizer is a unified library of state-of-the-art model optimization techniques such as quantization, pruning, distillati…☆534Updated this week