BorealisAI / efficient-vit-training
PyTorch code of "Training a Vision Transformer from scratch in less than 24 hours with 1 GPU" (HiTY workshop at Neurips 2022)
☆18Updated last year
Related projects ⓘ
Alternatives and complementary repositories for efficient-vit-training
- Official Pytorch Implementation of Self-emerging Token Labeling☆30Updated 7 months ago
- Pytorch Implementation of the paper: "Learning to (Learn at Test Time): RNNs with Expressive Hidden States"☆23Updated last week
- An efficient pytorch implementation of selective scan in one file, works with both cpu and gpu, with corresponding mathematical derivatio…☆71Updated 8 months ago
- Official code repository for paper: "ExPLoRA: Parameter-Efficient Extended Pre-training to Adapt Vision Transformers under Domain Shifts"☆24Updated last month
- A Contrastive Learning Boost from Intermediate Pre-Trained Representations☆36Updated last month
- Official PyTorch implementation of "LayerMerge: Neural Network Depth Compression through Layer Pruning and Merging" (ICML'24)☆27Updated 2 months ago
- SAM-CLIP module for use with Autodistill.☆12Updated 11 months ago
- This repository contains the pytorch code for our ISBI 2024 paper "ConvLoRA and AdaBN Based Domain Adaptation via Self-Training".☆54Updated last month
- Timm model explorer☆36Updated 7 months ago
- A simple but robust PyTorch implementation of RetNet from "Retentive Network: A Successor to Transformer for Large Language Models" (http…☆100Updated 11 months ago
- Implementation of a modular, high-performance, and simplistic mamba for high-speed applications☆33Updated last week
- Implementation of a Light Recurrent Unit in Pytorch☆46Updated last month
- Implementation of MoE Mamba from the paper: "MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts" in Pytorch and Ze…☆83Updated this week
- Code and pretrained models for the paper: "MatMamba: A Matryoshka State Space Model"☆50Updated last month
- Implementation of ViTaR: ViTAR: Vision Transformer with Any Resolution in PyTorch☆24Updated last week
- PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)☆64Updated last year
- The official repo for [TPAMI'23] "Vision Transformer with Quadrangle Attention"☆175Updated 7 months ago
- EdgeSAM model for use with Autodistill.☆25Updated 5 months ago
- ☆61Updated last month
- Mamba in Vision: A Comprehensive Survey of Techniques and Applications☆64Updated last month
- ☆49Updated last year
- [ECCV 2024] Isomorphic Pruning for Vision Models☆53Updated 3 months ago
- Introduce Mamba2 to Vision.☆91Updated 2 months ago
- ☆41Updated 7 months ago
- Code Implementation of EfficientVMamba☆184Updated 6 months ago
- [NeurIPS2022] This is the official implementation of the paper "Expediting Large-Scale Vision Transformer for Dense Prediction without Fi…☆82Updated last year
- My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing o…☆41Updated 11 months ago
- Data-Efficient Multimodal Fusion on a Single GPU☆47Updated 6 months ago
- Explorations into improving ViTArc with Slot Attention☆36Updated 3 weeks ago
- ML-Mamba: Efficient Multi-Modal Large Language Model Utilizing Mamba-2☆51Updated 3 weeks ago