Tony-Y / pytorch_warmupLinks
Learning Rate Warmup in PyTorch
☆414Updated 5 months ago
Alternatives and similar repositories for pytorch_warmup
Users that are interested in pytorch_warmup are comparing it to the libraries listed below
Sorting:
- Tiny PyTorch library for maintaining a moving average of a collection of parameters.☆439Updated last year
- ☆466Updated 2 years ago
- Gradually-Warmup Learning Rate Scheduler for PyTorch☆993Updated last year
- Implementation of ConvMixer for "Patches Are All You Need? 🤷"☆1,078Updated 3 years ago
- An All-MLP solution for Vision, from Google AI☆1,053Updated 4 months ago
- A PyTorch implementation of the 1d and 2d Sinusoidal positional encoding/embedding.☆260Updated 5 years ago
- Implementation of Transformer in Transformer, pixel level attention paired with patch level attention for image classification, in Pytorc…☆309Updated 3 years ago
- (ICLR 2022 Spotlight) Official PyTorch implementation of "How Do Vision Transformers Work?"☆821Updated 3 years ago
- Compute CNN receptive field size in pytorch in one line☆365Updated last year
- A Pytorch-Lightning implementation of self-supervised algorithms☆545Updated 3 years ago
- This is an official implementation for "Self-Supervised Learning with Swin Transformers".☆665Updated 4 years ago
- A PyTorch implementation of "CoAtNet: Marrying Convolution and Attention for All Data Sizes"☆392Updated 4 years ago
- Code for the Convolutional Vision Transformer (ConViT)☆470Updated 4 years ago
- NFNets and Adaptive Gradient Clipping for SGD implemented in PyTorch. Find explanation at tourdeml.github.io/blog/☆349Updated last year
- Escaping the Big Data Paradigm with Compact Transformers, 2021 (Train your Vision Transformers in 30 mins on CIFAR-10 with a single GPU!)☆538Updated last year
- Unofficial PyTorch implementation of "Meta Pseudo Labels"☆390Updated last year
- Implementation of Axial attention - attending to multi-dimensional data efficiently☆391Updated 4 years ago
- Unofficial implementation of MLP-Mixer: An all-MLP Architecture for Vision☆217Updated 4 years ago
- 🛠 Toolbox to extend PyTorch functionalities☆419Updated last year
- Is the attention layer even necessary? (https://arxiv.org/abs/2105.02723)☆483Updated 4 years ago
- An (unofficial) implementation of Focal Loss, as described in the RetinaNet paper, generalized to the multi-class case.☆239Updated last year
- Implementation of Linformer for Pytorch☆302Updated last year
- PyTorch implementation of Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning☆501Updated 3 years ago
- An implementation of the efficient attention module.☆326Updated 5 years ago
- A PyTorch Implementation of Focal Loss.☆990Updated 6 years ago
- Self-supervised vIsion Transformer (SiT)☆337Updated 2 years ago
- Unofficial PyTorch Reimplementation of RandAugment.☆638Updated 2 years ago
- A simple way to keep track of an Exponential Moving Average (EMA) version of your Pytorch model☆624Updated last year
- An implementation of 1D, 2D, and 3D positional encoding in Pytorch and TensorFlow☆611Updated last year
- Unofficial implementation of Google's FNet: Mixing Tokens with Fourier Transforms☆260Updated 4 years ago