yuanwei2019 / EAdam-optimizerLinks
Some improvements on Adam
☆28Updated 5 years ago
Alternatives and similar repositories for EAdam-optimizer
Users that are interested in EAdam-optimizer are comparing it to the libraries listed below
Sorting:
- [ICML 2020] code for the flooding regularizer proposed in "Do We Need Zero Training Loss After Achieving Zero Training Error?"☆95Updated 2 years ago
- A pytorch realization of adafactor (https://arxiv.org/pdf/1804.04235.pdf )☆25Updated 6 years ago
- Transformers are Graph Neural Networks!☆54Updated 4 years ago
- PyTorch implementation of FNet: Mixing Tokens with Fourier transforms☆28Updated 4 years ago
- PyTorch Examples repo for "ReZero is All You Need: Fast Convergence at Large Depth"☆62Updated last year
- The official implementation of paper "Drop-Activation: Implicit Parameter Reduction and Harmonious Regularization".☆10Updated 6 years ago
- Official code repository of the paper Linear Transformers Are Secretly Fast Weight Programmers.☆110Updated 4 years ago
- diffGrad: An Optimization Method for Convolutional Neural Networks☆54Updated 3 years ago
- [EMNLP'19] Summary for Transformer Understanding☆53Updated 5 years ago
- Interpolation between Residual and Non-Residual Networks, ICML 2020. https://arxiv.org/abs/2006.05749☆26Updated 5 years ago
- PyTorch Codes for Haar Graph Pooling☆11Updated 2 years ago
- NeurIPS 2022: Tree Mover’s Distance: Bridging Graph Metrics and Stability of Graph Neural Networks☆37Updated 2 years ago
- The implementation of paper ''Efficient Attention Network: Accelerate Attention by Searching Where to Plug''.☆20Updated 2 years ago
- Bootstrapping loss function implementation in pytorch☆36Updated 4 years ago
- TedNet: A Pytorch Toolkit for Tensor Decomposition Networks☆96Updated 3 years ago
- Skyformer: Remodel Self-Attention with Gaussian Kernel and Nystr\"om Method (NeurIPS 2021)☆63Updated 3 years ago
- ☆33Updated 4 years ago
- Unofficial PyTorch implementation of Fastformer based on paper "Fastformer: Additive Attention Can Be All You Need"."☆133Updated 4 years ago
- Graph neural network message passing reframed as a Transformer with local attention☆70Updated 2 years ago
- AdaX: Adaptive Gradient Descent with Exponential Long Term Momery☆34Updated 5 years ago
- Apollo: An Adaptive Parameter-wise Diagonal Quasi-Newton Method for Nonconvex Stochastic Optimization☆182Updated 4 years ago
- [ICML 2020] code for "PowerNorm: Rethinking Batch Normalization in Transformers" https://arxiv.org/abs/2003.07845☆120Updated 4 years ago
- A simple implementation of a deep linear Pytorch module☆21Updated 5 years ago
- PyTorch implementation of "MLP-Mixer: An all-MLP Architecture for Vision" Tolstikhin et al. (2021)☆31Updated 4 years ago
- A quick walk-through of the innards of LSTMs and a naive implementation of the Mogrifier LSTM paper in PyTorch☆78Updated 5 years ago
- ☆47Updated 4 years ago
- MODALS: Modality-agnostic Automated Data Augmentation in the Latent Space☆41Updated 4 years ago
- Implementation of Lie Transformer, Equivariant Self-Attention, in Pytorch☆96Updated 4 years ago
- ☆41Updated 4 years ago
- Implementation of Online Label Smoothing in PyTorch☆95Updated 3 years ago