LukasHedegaard / continual-transformersLinks
Official Pytorch Implementation for "Continual Transformers: Redundancy-Free Attention for Online Inference" [ICLR 2023]
☆28Updated 2 years ago
Alternatives and similar repositories for continual-transformers
Users that are interested in continual-transformers are comparing it to the libraries listed below
Sorting:
- Unofficial PyTorch implementation of the paper "cosFormer: Rethinking Softmax In Attention".☆44Updated 4 years ago
- Implementation of Memformer, a Memory-augmented Transformer, in Pytorch☆126Updated 5 years ago
- A Python library for Continual Inference Networks in PyTorch☆54Updated 9 months ago
- ☆73Updated 4 years ago
- Implementation of Multistream Transformers in Pytorch☆54Updated 4 years ago
- This is the public github for our paper "Transformer with a Mixture of Gaussian Keys"☆28Updated 3 years ago
- A variant of Transformer-XL where the memory is updated not with a queue, but with attention☆49Updated 5 years ago
- Implementation of Long-Short Transformer, combining local and global inductive biases for attention over long sequences, in Pytorch☆120Updated 4 years ago
- VIsually-Pivoted Audio and(N) Text☆22Updated 3 years ago
- EgoCom: A Multi-person Multi-modal Egocentric Communications Dataset☆59Updated 5 years ago
- Code for the AVLnet (Interspeech 2021) and Cascaded Multilingual (Interspeech 2021) papers.☆53Updated 3 years ago
- Generalized cross-modal NNs; new audiovisual benchmark (IEEE TNNLS 2019)☆30Updated 5 years ago
- Code for the paper PermuteFormer☆42Updated 4 years ago
- A minimal pytorch package implementing a gradient reversal layer.☆158Updated last year
- Continuous Augmented Positional Embeddings (CAPE) implementation for PyTorch☆42Updated 3 years ago
- Evaluation script for VoxMovies dataset in PyTorch☆23Updated 2 years ago
- Implementation of "compositional attention" from MILA, a multi-head attention variant that is reframed as a two-step attention process wi…☆51Updated 3 years ago
- Implementation of Cross Transformer for spatially-aware few-shot transfer, in Pytorch☆54Updated 4 years ago
- Implementation of OmniNet, Omnidirectional Representations from Transformers, in Pytorch☆59Updated 4 years ago
- [TMLR 2022] High-Modality Multimodal Transformer☆117Updated last year
- Code for the IEEE Signal Processing Letters 2022 paper "UAVM: Towards Unifying Audio and Visual Models".☆57Updated 2 years ago
- Official PyTorch implementation of Time-aware Large Kernel (TaLK) Convolutions (ICML 2020)☆29Updated 5 years ago
- ☆31Updated 4 years ago
- Relative Positional Encoding for Transformers with Linear Complexity☆65Updated 3 years ago
- Representation learning for NLP @ JSALT19☆40Updated 5 years ago
- [AAAI 2023 (Oral)] CrissCross: Self-Supervised Audio-Visual Representation Learning with Relaxed Cross-Modal Synchronicity☆25Updated 2 years ago
- Skyformer: Remodel Self-Attention with Gaussian Kernel and Nystr\"om Method (NeurIPS 2021)☆63Updated 3 years ago
- custom pytorch implementation of MoCo v3☆46Updated 4 years ago
- ☆23Updated 5 years ago
- Pytorch Implementation for "Preserving Linear Separability in Continual Learning by Backward Feature Projection" (CVPR 2023)☆18Updated 2 years ago