neuralmagic / docs
Top-level directory for documentation and general content
☆120Updated 4 months ago
Alternatives and similar repositories for docs:
Users that are interested in docs are comparing it to the libraries listed below
- Neural network model repository for highly sparse and sparse-quantized models with matching sparsification recipes☆382Updated 9 months ago
- ML model optimization product to accelerate inference.☆326Updated last year
- Libraries for applying sparsification recipes to neural networks with a few lines of code, enabling faster and smaller models☆2,124Updated 8 months ago
- Sparsity-aware deep learning inference runtime for CPUs☆3,133Updated 9 months ago
- PyTorch library to facilitate development and standardized evaluation of neural network pruning methods.☆429Updated last year
- A research library for pytorch-based neural network pruning, compression, and more.☆160Updated 2 years ago
- Prune a model while finetuning or training.☆402Updated 2 years ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆262Updated 6 months ago
- End-to-end training of sparse deep neural networks with little-to-no performance loss.☆321Updated 2 years ago
- A library for researching neural networks compression and acceleration methods.☆139Updated 7 months ago
- Template repository to build PyTorch projects from source on any version of PyTorch/CUDA/cuDNN.☆719Updated 3 months ago
- FasterAI: Prune and Distill your models with FastAI and PyTorch☆247Updated 2 weeks ago
- Neural Network Compression Framework for enhanced OpenVINO™ inference☆998Updated this week
- Accelerate your Neural Architecture Search (NAS) through fast, reproducible and modular research.☆475Updated 5 months ago
- SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX R…☆2,376Updated this week
- MONeT framework for reducing memory consumption of DNN training☆173Updated 3 years ago
- ☆876Updated last year
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆459Updated this week
- A Python-level JIT compiler designed to make unmodified PyTorch programs faster.☆1,038Updated last year
- Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackab…☆1,565Updated last year
- TorchX is a universal job launcher for PyTorch applications. TorchX is designed to have fast iteration time for training/research and sup…☆358Updated this week
- ☆202Updated 2 years ago
- A GPU performance profiling tool for PyTorch models☆506Updated 3 years ago
- ☆20Updated 9 months ago
- A repository in preparation for open-sourcing lottery ticket hypothesis code.☆630Updated 2 years ago
- ☆202Updated 3 years ago
- [NeurIPS 2020] MCUNet: Tiny Deep Learning on IoT Devices; [NeurIPS 2021] MCUNetV2: Memory-Efficient Patch-based Inference for Tiny Deep L…☆556Updated last year
- Pax is a Jax-based machine learning framework for training large scale models. Pax allows for advanced and fully configurable experimenta…☆488Updated this week
- TorchBench is a collection of open source benchmarks used to evaluate PyTorch performance.☆933Updated this week
- Repository to perform simple pruning experiments on neural networks☆18Updated 4 years ago