Mixture of Attention Heads
☆52Oct 10, 2022Updated 3 years ago
Alternatives and similar repositories for MoA
Users that are interested in MoA are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- sigma-MoE layer☆21Jan 5, 2024Updated 2 years ago
- [ICLR 2023] "Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable Transformers" by Tianlong Chen*, Zhenyu Zhang*, Ajay Jaiswal…☆56Feb 28, 2023Updated 3 years ago
- Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"☆101Sep 30, 2024Updated last year
- ☆14Nov 20, 2022Updated 3 years ago
- [ACL 2023 Findings] Emergent Modularity in Pre-trained Transformers☆26Jun 7, 2023Updated 2 years ago
- ☆11Jan 18, 2025Updated last year
- ☆11Sep 7, 2024Updated last year
- ☆17Jun 11, 2025Updated 9 months ago
- Reranking-based dependency parsing with inside-outside recursive neural network☆21Oct 11, 2014Updated 11 years ago
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆39Jun 11, 2025Updated 9 months ago
- Code for the ACL-2022 paper "StableMoE: Stable Routing Strategy for Mixture of Experts"☆51Jul 17, 2022Updated 3 years ago
- [ACL 2019/AACL 2020] Second-Order Syntactic/Semantic Dependency Parsing With Mean Field Variational Inference (PyTorch)☆14Oct 22, 2020Updated 5 years ago
- Fine-Tuning Pre-trained Transformers into Decaying Fast Weights☆19Oct 9, 2022Updated 3 years ago
- The official repository for our paper "The Neural Data Router: Adaptive Control Flow in Transformers Improves Systematic Generalization".☆34Jun 11, 2025Updated 9 months ago
- ☆19Nov 5, 2024Updated last year
- [ICML 2024] When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models☆35Jun 12, 2024Updated last year
- Triton Implementation of HyperAttention Algorithm☆48Dec 11, 2023Updated 2 years ago
- (ACM MM24) This is the offical repository of GIST: Improving Parameter Efficient Fine Tuning via Knowledge Interaction.☆11Jan 28, 2024Updated 2 years ago
- Towards Real-World Writing Assistance: A Chinese Character Checking Benchmark with Faked and Misspelled Characters☆16May 30, 2024Updated last year
- Implementation of Pix2Seq in PyTorch☆10Feb 3, 2022Updated 4 years ago
- ☆125Feb 4, 2026Updated last month
- Official implementation of "A Backpack Full of Skills: Egocentric Video Understanding with Diverse Task Perspectives", accepted at CVPR 2…☆24Jun 13, 2024Updated last year
- ☆11Jul 30, 2016Updated 9 years ago
- The source code of "Merging Experts into One: Improving Computational Efficiency of Mixture of Experts (EMNLP 2023)":☆44Feb 28, 2026Updated 3 weeks ago
- [EMNLP 2023] Official implementation of the algorithm ETSC: Exact Toeplitz-to-SSM Conversion our EMNLP 2023 paper - Accelerating Toeplitz…☆14Oct 17, 2023Updated 2 years ago
- ☆19Oct 31, 2022Updated 3 years ago
- [EMNLP'23] Code for "Non-autoregressive Text Editing with Copy-aware Latent Alignments".☆20Oct 17, 2023Updated 2 years ago
- [NeurIPS 2024] EvolveDirector: Approaching Advanced Text-to-Image Generation with Large Vision-Language Models.☆52Oct 14, 2024Updated last year
- This PyTorch package implements MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation (NAACL 2022).☆114May 2, 2022Updated 3 years ago
- Repo for paper "MUSEG: Reinforcing Video Temporal Understanding via Timestamp-Aware Multi-Segment Grounding".☆39Jun 9, 2025Updated 9 months ago
- User-friendly implementation of the Mixture-of-Sparse-Attention (MoSA). MoSA selects distinct tokens for each head with expert choice rou…☆28May 3, 2025Updated 10 months ago
- ☆45Apr 30, 2018Updated 7 years ago
- source code of COLING2020 "Second-Order Unsupervised Neural Dependency Parsing"☆16Oct 24, 2022Updated 3 years ago
- Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"☆59Jan 12, 2023Updated 3 years ago
- Efficient PScan implementation in PyTorch☆17Jan 2, 2024Updated 2 years ago
- ☆45Sep 26, 2021Updated 4 years ago
- Video Games Dataset for Multi-Document Summarization☆19Sep 20, 2025Updated 6 months ago
- How certain is your transformer?☆25Apr 25, 2021Updated 4 years ago
- LEO: A powerful Hybrid Multimodal LLM☆20Jan 18, 2025Updated last year