utsaslab / MONeT
MONeT framework for reducing memory consumption of DNN training
☆174Updated 3 years ago
Alternatives and similar repositories for MONeT:
Users that are interested in MONeT are comparing it to the libraries listed below
- PyTorch implementation of L2L execution algorithm☆107Updated 2 years ago
- PyProf2: PyTorch Profiling tool☆82Updated 4 years ago
- ☆109Updated 3 years ago
- Train ImageNet in 18 minutes on AWS☆128Updated 11 months ago
- Distributed, mixed-precision training with PyTorch☆89Updated 4 years ago
- Slicing a PyTorch Tensor Into Parallel Shards☆298Updated 3 years ago
- A GPU performance profiling tool for PyTorch models☆501Updated 3 years ago
- Example code showing how to use Nvidia DALI in pytorch, with fallback to torchvision. Contains a few differences to the official Nvidia …☆197Updated 5 years ago
- Fairring (FAIR + Herring) is a plug-in for PyTorch that provides a process group for distributed training that outperforms NCCL at large …☆64Updated 2 years ago
- Programmable Neural Network Compression☆148Updated 2 years ago
- "Layer-wise Adaptive Rate Scaling" in PyTorch☆86Updated 4 years ago
- Labels and other data for the paper "Are we done with ImageNet?"☆189Updated 3 years ago
- PyTorch layer-by-layer model profiler☆606Updated 3 years ago
- [Prototype] Tools for the concurrent manipulation of variably sized Tensors.☆253Updated 2 years ago
- Lightweight and Parallel Deep Learning Framework☆264Updated 2 years ago
- Experimental ground for optimizing memory of pytorch models☆364Updated 6 years ago
- Training neural networks in TensorFlow 2.0 with 5x less memory☆130Updated 2 years ago
- A Re-implementation of Fixed-update Initialization☆152Updated 5 years ago
- Block-sparse primitives for PyTorch☆153Updated 3 years ago
- This repository contains the results and code for the MLPerf™ Training v0.7 benchmark.☆56Updated last year
- [ICLR 2020] Drawing Early-Bird Tickets: Toward More Efficient Training of Deep Networks☆137Updated 4 years ago
- ☆57Updated 2 years ago
- ☆62Updated 4 years ago
- ☆143Updated last year
- DiffQ performs differentiable quantization using pseudo quantization noise. It can automatically tune the number of bits used per weight …☆235Updated last year
- Code for paper "SWALP: Stochastic Weight Averaging forLow-Precision Training".☆62Updated 5 years ago
- On Network Design Spaces for Visual Recognition☆94Updated 4 years ago
- Seamless analysis of your PyTorch models (RAM usage, FLOPs, MACs, receptive field, etc.)☆216Updated last week
- Estimate/count FLOPS for a given neural network using pytorch☆304Updated 2 years ago
- Accelerate training by storing parameters in one contiguous chunk of memory.☆292Updated 4 years ago