SHI-Labs / NATTENLinks
Neighborhood Attention Extension. Bringing attention to a neighborhood near you!
☆528Updated this week
Alternatives and similar repositories for NATTEN
Users that are interested in NATTEN are comparing it to the libraries listed below
Sorting:
- [ECCV 2024] Official PyTorch implementation of RoPE-ViT "Rotary Position Embedding for Vision Transformer"☆335Updated 6 months ago
- Neighborhood Attention Transformer, arxiv 2022 / CVPR 2023. Dilated Neighborhood Attention Transformer, arxiv 2022☆1,122Updated last year
- Causal depthwise conv1d in CUDA, with a PyTorch interface☆494Updated 3 weeks ago
- This repo contains the code for 1D tokenizer and generator☆918Updated 3 months ago
- [ICLR'25 Oral] Representation Alignment for Generation: Training Diffusion Transformers Is Easier Than You Think☆1,126Updated 3 months ago
- [ICLR 2024] Official PyTorch implementation of FasterViT: Fast Vision Transformers with Hierarchical Attention☆855Updated 3 months ago
- Official PyTorch Implementation of "SiT: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers"☆873Updated last year
- Masked Diffusion Transformer is the SOTA for image synthesis. (ICCV 2023)☆568Updated last year
- [ICLR 2025 Spotlight] Vision-RWKV: Efficient and Scalable Visual Perception with RWKV-Like Architectures☆472Updated 4 months ago
- Implementation of MagViT2 Tokenizer in Pytorch☆608Updated 5 months ago
- [NeurIPS 2024] Official implementation of "Faster Diffusion: Rethinking the Role of UNet Encoder in Diffusion Models"☆336Updated 3 months ago
- [ECCV 2024] Official Repository for DiffiT: Diffusion Vision Transformers for Image Generation☆495Updated 7 months ago
- Helpful tools and examples for working with flex-attention☆831Updated 2 weeks ago
- A PyTorch implementation of the paper "ZigMa: A DiT-Style Mamba-based Diffusion Model" (ECCV 2024)☆318Updated 3 months ago
- [CVPR 2025 Oral] Reconstruction vs. Generation: Taming Optimization Dilemma in Latent Diffusion Models☆949Updated last week
- Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attention Does Not Need O(n²) Memory"☆379Updated last year
- Implementation of Rotary Embeddings, from the Roformer paper, in Pytorch☆694Updated 6 months ago
- SEED-Voken: A Series of Powerful Visual Tokenizers☆897Updated 4 months ago
- Code for Fast Training of Diffusion Models with Masked Transformers☆403Updated last year
- Implementation of Deformable Attention in Pytorch from the paper "Vision Transformer with Deformable Attention"☆344Updated 4 months ago
- A simple way to keep track of an Exponential Moving Average (EMA) version of your Pytorch model☆592Updated 6 months ago
- Implementation of a single layer of the MMDiT, proposed in Stable Diffusion 3, in Pytorch☆369Updated 5 months ago
- A method to increase the speed and lower the memory footprint of existing vision transformers.☆1,064Updated last year
- MetaFormer Baselines for Vision (TPAMI 2024)☆472Updated last year
- Scaling Diffusion Transformers with Mixture of Experts☆339Updated 9 months ago
- EDM2 and Autoguidance -- Official PyTorch implementation☆729Updated 6 months ago
- [ECCV 2022] Official repository for "MaxViT: Multi-Axis Vision Transformer". SOTA foundation models for classification, detection, segmen…☆475Updated 2 years ago
- [ECCV2024] VideoMamba: State Space Model for Efficient Video Understanding☆963Updated 11 months ago
- [ICLR2025] Halton Scheduler for Masked Generative Image Transformer☆239Updated last month
- An efficient pytorch implementation of selective scan in one file, works with both cpu and gpu, with corresponding mathematical derivatio…☆91Updated last year