LiyuanLucasLiu / RAdam
On the Variance of the Adaptive Learning Rate and Beyond
☆2,546Updated 3 years ago
Alternatives and similar repositories for RAdam:
Users that are interested in RAdam are comparing it to the libraries listed below
- Ranger - a synergistic optimizer using RAdam (Rectified Adam), Gradient Centralization and LookAhead in one codebase☆1,196Updated last year
- An optimizer that trains as fast as Adam and as good as SGD.☆2,910Updated last year
- Model summary in PyTorch similar to `model.summary()` in Keras☆4,033Updated last year
- Codebase for Image Classification Research, written in PyTorch.☆2,150Updated last year
- Neural network graphs and training metrics for PyTorch, Tensorflow, and Keras.☆1,818Updated last year
- torch-optimizer -- collection of optimizers for Pytorch☆3,092Updated last year
- A lightweight library for PyTorch training tools and utilities☆1,683Updated this week
- Official Implementation of 'Fast AutoAugment' in PyTorch.☆1,602Updated 3 years ago
- PyTorch implementation of "Efficient Neural Architecture Search via Parameters Sharing"☆2,710Updated last year
- Implementation of LambdaNetworks, a new approach to image recognition that reaches SOTA with less compute☆1,531Updated 4 years ago
- Debug PyTorch code using PySnooper☆799Updated 3 years ago
- Unsupervised Data Augmentation (UDA)☆2,188Updated 3 years ago
- tensorboard for pytorch (and chainer, mxnet, numpy, ...)☆7,919Updated 2 months ago
- Train AI models efficiently on medical images using any framework☆1,871Updated 9 months ago
- pip install antialiased-cnns to improve stability and accuracy☆1,667Updated 11 months ago
- Pretrained EfficientNet, EfficientNet-Lite, MixNet, MobileNetV3 / V2, MNASNet A1 and B1, FBNet, Single-Path NAS☆1,568Updated 9 months ago
- High-level library to help with training and evaluating neural networks in PyTorch flexibly and transparently.☆4,598Updated this week
- Neural Network Distiller by Intel AI Lab: a Python package for neural network compression research. https://intellabs.github.io/distille…☆4,376Updated last year
- Stochastic Weight Averaging in PyTorch☆969Updated 3 years ago
- My best practice of training large dataset using PyTorch.☆1,092Updated 10 months ago
- ☆1,137Updated last year
- Differentiable architecture search for convolutional and recurrent networks☆3,943Updated 4 years ago
- A small package to create visualizations of PyTorch execution graphs☆3,320Updated 2 months ago
- Official Repository for "Mish: A Self Regularized Non-Monotonic Neural Activation Function" [BMVC 2020]☆1,296Updated last week
- A PyTorch Extension: Tools for easy mixed precision and distributed training in Pytorch☆8,585Updated this week
- Unofficial implementation of the ImageNet, CIFAR 10 and SVHN Augmentation Policies learned by AutoAugment using pillow☆1,480Updated last year
- ResNeSt: Split-Attention Networks☆3,246Updated 2 years ago
- Repository for NeurIPS 2020 Spotlight "AdaBelief Optimizer: Adapting stepsizes by the belief in observed gradients"☆1,060Updated 7 months ago
- Examples of using sparse attention, as in "Generating Long Sequences with Sparse Transformers"☆1,560Updated 4 years ago
- This repository reproduces the results of the paper: "Fixing the train-test resolution discrepancy" https://arxiv.org/abs/1906.06423☆1,039Updated 3 years ago