BobMcDear / vit-pytorch
PyTorch implementation of the vision transformer
☆19Updated last year
Related projects ⓘ
Alternatives and complementary repositories for vit-pytorch
- Neural network from scratch in CUDA/C++☆69Updated last year
- PyTorch implementation of SimSiam☆8Updated last year
- PyTorch implementation of popular attention mechanisms in vision☆15Updated last year
- A Pytorch implementation of Logic Tensor Networks☆8Updated 5 years ago
- A set of of fundamental operations and deep learning models using JAX☆13Updated 3 years ago
- Flax Image Models - State-of-the-art pre-trained vision backbones for Flax.☆17Updated last year
- This is a simple torch implementation of the high performance Multi-Query Attention☆15Updated last year
- NeurIPS2021: Vision Transformer Paper Collection☆9Updated 2 years ago
- ☆16Updated last year
- PyTorch implementation of FNet: Mixing Tokens with Fourier transforms☆25Updated 3 years ago
- Factorized Neural Layers☆27Updated last year
- Little article showing how to load pytorch's models with linear memory consumption☆34Updated 2 years ago
- Contains my experiments with the `big_vision` repo to train ViTs on ImageNet-1k.☆22Updated last year
- This repository hosts the code to port NumPy model weights of BiT-ResNets to TensorFlow SavedModel format.☆14Updated 2 years ago
- HGRN2: Gated Linear RNNs with State Expansion☆49Updated 3 months ago
- ⛰️ RockyML - A High-Performance Scientific Computing Framework for Non-smooth Machine Learning Problems☆19Updated last year
- Examples of using PyTorch hooks, as covered in my YouTube tutorial video.☆32Updated last year
- JAX/Flax implimentation of 'Attention Is All You Need' by Vaswani et al. (https://arxiv.org/abs/1706.03762)☆12Updated 3 years ago
- This is the public github for our paper "Transformer with a Mixture of Gaussian Keys"☆26Updated 2 years ago
- JAX implementation ViT-VQGAN☆55Updated 2 years ago
- Code for the PAPA paper☆27Updated 2 years ago
- Deep Learning Experiment Code.☆19Updated 3 months ago
- ☆51Updated 5 months ago
- Simply Numpy implementation of the FAVOR+ attention mechanism, https://teddykoker.com/2020/11/performers/☆37Updated 3 years ago
- Exploring an idea where one forgets about efficiency and carries out attention across each edge of the nodes (tokens)☆43Updated last month
- Skyformer: Remodel Self-Attention with Gaussian Kernel and Nystr\"om Method (NeurIPS 2021)☆59Updated 2 years ago
- Explorations into the recently proposed Taylor Series Linear Attention☆90Updated 3 months ago
- Transformers w/o Attention, based fully on MLPs☆89Updated 7 months ago
- Code for the paper PermuteFormer☆42Updated 3 years ago
- A Python package for generating concise, high-quality summaries of a probability distribution☆41Updated 3 weeks ago