neuralmagic / docsLinks
Top-level directory for documentation and general content
☆120Updated 7 months ago
Alternatives and similar repositories for docs
Users that are interested in docs are comparing it to the libraries listed below
Sorting:
- Neural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes☆387Updated 7 months ago
- Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models☆2,143Updated 7 months ago
- ML model optimization product to accelerate inference.☆325Updated 7 months ago
- Sparsity-aware deep learning inference runtime for CPUs☆3,158Updated 7 months ago
- A research library for pytorch-based neural network pruning, compression, and more.☆162Updated 3 years ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆1,115Updated this week
- Prune a model while finetuning or training.☆405Updated 3 years ago
- Accelerate your Neural Architecture Search (NAS) through fast, reproducible and modular research.☆484Updated last month
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,071Updated last year
- Model Compression Toolkit (MCT) is an open source project for neural network model optimization under efficient, constrained hardware. Th…☆431Updated this week
- PyTorch library to facilitate development and standardized evaluation of neural network pruning methods.☆432Updated 2 years ago
- FasterAI: Prune and Distill your models with FastAI and PyTorch☆252Updated this week
- Recipes are a standard, well supported set of blueprints for machine learning engineers to rapidly train models using the latest research…☆339Updated this week
- End-to-end training of sparse deep neural networks with little-to-no performance loss.☆333Updated 2 years ago
- GPU implementation of a fast generalized ANS (asymmetric numeral system) entropy encoder and decoder, with extensions for lossless compre…☆369Updated this week
- A Pytorch Knowledge Distillation library for benchmarking and extending works in the domains of Knowledge Distillation, Pruning, and Quan…☆653Updated 2 years ago
- Fast Block Sparse Matrices for Pytorch☆550Updated 4 years ago
- Repository to perform simple pruning experiments on neural networks☆18Updated 5 years ago
- Triton Model Analyzer is a CLI tool to help with better understanding of the compute and memory requirements of the Triton Inference Serv…☆502Updated this week
- Easily benchmark PyTorch model FLOPs, latency, throughput, allocated gpu memory and energy consumption☆109Updated 2 years ago
- Library for 8-bit optimizers and quantization routines.☆781Updated 3 years ago
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆528Updated this week
- An open-source efficient deep learning framework/compiler, written in python.☆739Updated 4 months ago
- Template repository to build PyTorch projects from source on any version of PyTorch/CUDA/cuDNN.☆722Updated last year
- ☆227Updated last year
- MLCommons Algorithmic Efficiency is a benchmark and competition measuring neural network training speedups due to algorithmic improvement…☆406Updated this week
- SOTA low-bit LLM quantization (INT8/FP8/MXFP8/INT4/MXFP4/NVFP4) & sparsity; leading model compression techniques on PyTorch, TensorFlow, …☆2,570Updated this week
- Convert ONNX models to PyTorch.☆720Updated 3 months ago
- Accelerate PyTorch models with ONNX Runtime☆368Updated last month
- Fast sparse deep learning on CPUs☆56Updated 3 years ago