Caiyun-AI / MUDDFormerLinks
☆83Updated 4 months ago
Alternatives and similar repositories for MUDDFormer
Users that are interested in MUDDFormer are comparing it to the libraries listed below
Sorting:
- [ICML 2025] Fourier Position Embedding: Enhancing Attention’s Periodic Extension for Length Generalization☆98Updated 3 months ago
- Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models☆328Updated 6 months ago
- ☆216Updated 7 months ago
- ☆197Updated last year
- The official GitHub page for the survey paper "A Survey of RWKV".☆30Updated 8 months ago
- A repository for DenseSSMs☆88Updated last year
- [EMNLP 2024] RWKV-CLIP: A Robust Vision-Language Representation Learner☆142Updated 4 months ago
- [COLM 2025] LoRI: Reducing Cross-Task Interference in Multi-Task Low-Rank Adaptation☆147Updated 2 months ago
- The official GitHub page for the survey paper "Discrete Tokenization for Multimodal LLMs: A Comprehensive Survey". And this paper is unde…☆63Updated last month
- ☆72Updated 7 months ago
- [ICML 2025 Oral] Mixture of Lookup Experts☆51Updated 4 months ago
- [ICLR 2025 Spotlight] Official Implementation for ToST (Token Statistics Transformer)☆118Updated 6 months ago
- TransMLA: Multi-Head Latent Attention Is All You Need☆356Updated 2 weeks ago
- ☆148Updated last year
- Parameter-Efficient Fine-Tuning for Foundation Models☆93Updated 5 months ago
- Official implementation of TPA: Tensor ProducT ATTenTion Transformer (T6) (https://arxiv.org/abs/2501.06425)☆386Updated this week
- ☆47Updated 3 months ago
- DeepSeek Native Sparse Attention pytorch implementation☆95Updated last month
- Official repository of InLine attention (NeurIPS 2024)☆55Updated 8 months ago
- Official code for our paper, "LoRA-Pro: Are Low-Rank Adapters Properly Optimized? "☆131Updated 5 months ago
- Lion and Adam optimization comparison☆64Updated 2 years ago
- A generalized framework for subspace tuning methods in parameter efficient fine-tuning.☆154Updated 2 months ago
- Implementation of Switch Transformers from the paper: "Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficien…☆120Updated last week
- Triton implement of bi-directional (non-causal) linear attention☆54Updated 7 months ago
- ☆116Updated last year
- qwen-nsa☆74Updated 5 months ago
- A Tight-fisted Optimizer☆50Updated 2 years ago
- State Space Models☆70Updated last year
- Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆106Updated this week
- The this is the official implementation of "DAPE: Data-Adaptive Positional Encoding for Length Extrapolation"☆39Updated 11 months ago