JingzhaoZhang / why-clipping-acceleratesLinks
A pytorch implementation for the LSTM experiments in the paper: Why Gradient Clipping Accelerates Training: A Theoretical Justification for Adaptivity
☆47Updated 5 years ago
Alternatives and similar repositories for why-clipping-accelerates
Users that are interested in why-clipping-accelerates are comparing it to the libraries listed below
Sorting:
- Reparameterize your PyTorch modules☆71Updated 5 years ago
- Implementation of Methods Proposed in Preventing Gradient Attenuation in Lipschitz Constrained Convolutional Networks (NeurIPS 2019)☆36Updated 5 years ago
- ☆47Updated 5 years ago
- This repository is no longer maintained. Check☆81Updated 5 years ago
- [NeurIPS'19] [PyTorch] Adaptive Regularization in NN☆68Updated 6 years ago
- Code base for SRSGD.☆28Updated 5 years ago
- The original code for the paper "How to train your MAML" along with a replication of the original "Model Agnostic Meta Learning" (MAML) p…☆41Updated 5 years ago
- Gradient Starvation: A Learning Proclivity in Neural Networks☆61Updated 5 years ago
- Code for "Supermasks in Superposition"☆125Updated 2 years ago
- [JMLR] TRADES + random smoothing for certifiable robustness☆14Updated 5 years ago
- ☆42Updated 2 years ago
- Implementation of Information Dropout☆39Updated 8 years ago
- PyTorch Examples repo for "ReZero is All You Need: Fast Convergence at Large Depth"☆62Updated last year
- PyTorch Implementations of Dropout Variants☆88Updated 8 years ago
- Code for Self-Tuning Networks (ICLR 2019) https://arxiv.org/abs/1903.03088☆61Updated 6 years ago
- ☆47Updated 6 years ago
- SGD and Ordered SGD codes for deep learning, SVM, and logistic regression☆36Updated 5 years ago
- An adaptive training algorithm for residual network☆17Updated 5 years ago
- [ICLR 2020] FSPool: Learning Set Representations with Featurewise Sort Pooling☆41Updated 2 years ago
- Geometric Certifications of Neural Nets☆42Updated 3 years ago
- ☆61Updated 2 years ago
- ICLR 2021, Fair Mixup: Fairness via Interpolation☆59Updated 4 years ago
- Tensorflow implementation of "Meta Dropout: Learning to Perturb Latent Features for Generalization" (ICLR 2020)☆27Updated 5 years ago
- A collection of Gradient-Based Meta-Learning Algorithms with pytorch☆65Updated 6 years ago
- [ICML 2020] code for "PowerNorm: Rethinking Batch Normalization in Transformers" https://arxiv.org/abs/2003.07845☆120Updated 4 years ago
- rich posterior approximations and anomaly detection☆20Updated 6 years ago
- Implementation of the models and datasets used in "An Information-theoretic Approach to Distribution Shifts"☆25Updated 4 years ago
- MTAdam: Automatic Balancing of Multiple Training Loss Terms☆36Updated 5 years ago
- Evaluating AlexNet features at various depths☆40Updated 5 years ago
- Low-variance, efficient and unbiased gradient estimation for optimizing models with binary latent variables. (ICLR 2019)☆27Updated 6 years ago