lucidrains / bidirectional-cross-attentionLinks
A simple cross attention that updates both the source and target in one step
☆173Updated last year
Alternatives and similar repositories for bidirectional-cross-attention
Users that are interested in bidirectional-cross-attention are comparing it to the libraries listed below
Sorting:
- Implementation of CrossViT: Cross-Attention Multi-Scale Vision Transformer for Image Classification☆201Updated 4 years ago
- Implementation of Zorro, Masked Multimodal Transformer, in Pytorch☆97Updated last year
- [NeurIPS 2023, Spotlight] Rank-N-Contrast: Learning Continuous Representations for Regression☆114Updated last year
- Learnable Fourier Features for Multi-Dimensional Spatial Positional Encoding☆51Updated 8 months ago
- [NeurIPS 2023] Factorized Contrastive Learning: Going Beyond Multi-view Redundancy☆69Updated last year
- Official Open Source code for "Masked Autoencoders As Spatiotemporal Learners"☆341Updated 7 months ago
- [T-PAMI] A curated list of self-supervised multimodal learning resources.☆261Updated 10 months ago
- MultiMAE: Multi-modal Multi-task Masked Autoencoders, ECCV 2022☆581Updated 2 years ago
- ☆158Updated last week
- A practical implementation of GradNorm, Gradient Normalization for Adaptive Loss Balancing, in Pytorch☆97Updated last year
- Code repository of the paper "Modelling Long Range Dependencies in ND: From Task-Specific to a General Purpose CNN" https://arxiv.org/abs…☆184Updated last month
- [ICLR 2022] Official implementation of cosformer-attention in cosFormer: Rethinking Softmax in Attention☆193Updated 2 years ago
- Implementation of Transformer in Transformer, pixel level attention paired with patch level attention for image classification, in Pytorc…☆305Updated 3 years ago
- Official implementation of "Hydra: Bidirectional State Space Models Through Generalized Matrix Mixers"☆138Updated 4 months ago
- Official PyTorch Implementation of "The Hidden Attention of Mamba Models"☆223Updated last year
- Official PyTorch implementation for the paper "CARD: Classification and Regression Diffusion Models"☆228Updated 2 years ago
- This repository is an implementation for the loss function proposed in https://arxiv.org/pdf/2110.06848.pdf.☆115Updated 3 years ago
- Transformers w/o Attention, based fully on MLPs☆93Updated last year
- The official repo for [TPAMI'23] "Vision Transformer with Quadrangle Attention"☆216Updated last year
- Pytorch implementation of Swin MAE https://arxiv.org/abs/2212.13805☆88Updated last month
- [CVPR'23 & TPAMI'25] Hard Patches Mining for Masked Image Modeling☆96Updated 2 months ago
- ConvMAE: Masked Convolution Meets Masked Autoencoders☆507Updated 2 years ago
- An implementation of local windowed attention for language modeling☆454Updated 5 months ago
- Implementation of Nyström Self-attention, from the paper Nyströmformer☆135Updated 3 months ago
- Ofiicial Implementation for Mamba-ND: Selective State Space Modeling for Multi-Dimensional Data☆60Updated 11 months ago
- Implementation of Axial attention - attending to multi-dimensional data efficiently☆384Updated 3 years ago
- Official Implementation of the CrossMAE paper: Rethinking Patch Dependence for Masked Autoencoders☆114Updated 2 months ago
- (ICLR 2023) Official PyTorch implementation of "What Do Self-Supervised Vision Transformers Learn?"☆112Updated last year
- [NeurIPS 2021] Official codes for "Efficient Training of Visual Transformers with Small Datasets".☆144Updated 5 months ago
- Official implementation of "A simple, efficient and scalable contrastive masked autoencoder for learning visual representations".☆32Updated 2 years ago