taki0112 / RAdam-TensorflowLinks
Simple Tensorflow implementation of "On The Variance Of The Adaptive Learning Rate And Beyond"
☆97Updated 5 years ago
Alternatives and similar repositories for RAdam-Tensorflow
Users that are interested in RAdam-Tensorflow are comparing it to the libraries listed below
Sorting:
- Simple Tensorflow implementation of "Adaptive Gradient Methods with Dynamic Bound of Learning Rate" (ICLR 2019)☆150Updated 6 years ago
- A simpler version of the self-attention layer from SAGAN, and some image classification results.☆214Updated 6 years ago
- Corrupted labels and label smoothing☆129Updated 8 years ago
- Code for paper: "Support Vector Machines, Wasserstein's distance and gradient-penalty GANs maximize a margin"☆179Updated 5 years ago
- Implementation of Rectified Adam in Keras☆70Updated 6 years ago
- An optimizer that trains as fast as Adam and as good as SGD in Tensorflow☆46Updated 6 years ago
- Python way to Read/Write TFRecords☆65Updated 7 years ago
- Simple Tensorflow implementation of "Partial Convolution based Padding" (partialconv)☆91Updated 7 years ago
- TensorFlow implementations of Wasserstein GAN with Gradient Penalty (WGAN-GP), Least Squares GAN (LSGAN), GANs with the hinge loss.☆44Updated 6 years ago
- Mish Deep Learning Activation Function for PyTorch / FastAI☆161Updated 5 years ago
- Code for reproducing results of the paper "Layer rotation: a surprisingly powerful indicator of generalization in deep networks?"☆50Updated 6 years ago
- Simple Tensorflow implementation of "On the Convergence of Adam and Beyond" (ICLR 2018)☆104Updated 6 years ago
- Implementation of tools to control and monitor layer rotation in different DL libraries☆40Updated 6 years ago
- Switch Normalization implementation for Keras 2+☆30Updated 7 years ago
- tunz's CUDA pytorch operator (MaskedSoftmax)☆75Updated 6 years ago
- Compare outputs between layers written in Tensorflow and layers written in Pytorch☆72Updated 7 years ago
- AdamW optimizer for Keras☆116Updated 6 years ago
- Keras implementation of Padam from "Closing the Generalization Gap of Adaptive Gradient Methods in Training Deep Neural Networks"☆17Updated 7 years ago
- Keras/TF implementation of AdamW, SGDW, NadamW, Warm Restarts, and Learning Rate multipliers☆169Updated 4 years ago
- "Learning Rate Dropout" in PyTorch☆34Updated 6 years ago
- Keras implementation of Octave Convolutions☆53Updated 6 years ago
- Implementation of the LAMB optimizer for Keras from the paper "Reducing BERT Pre-Training Time from 3 Days to 76 Minutes"☆75Updated 6 years ago
- Implementation of "Learning with Random Learning Rates" in PyTorch.☆102Updated 6 years ago
- Experiments with Adam/AdamW/amsgrad☆201Updated 7 years ago
- A set of simple examples ported from PyTorch for Tensorflow Eager Execution☆73Updated 7 years ago
- keras implementation of AdamW from Fixing Weight Decay Regularization in Adam (https://arxiv.org/abs/1711.05101)☆71Updated 7 years ago
- Simple Tensorflow implementation of SphereGAN (CVPR 2019 Oral)☆56Updated 6 years ago
- Knowledge Distillation Toolkit☆89Updated 5 years ago
- Torchélie is a set of utility functions, layers, losses, models, trainers and other things for PyTorch.☆110Updated 3 weeks ago
- Octave convolution☆34Updated 3 years ago