UNITES-Lab / Lingual-SMoELinks
[ICLR 2024] Code for the paper "Sparse MoE with Language-Guided Routing for Multilingual Machine Translation"
☆9Updated last year
Alternatives and similar repositories for Lingual-SMoE
Users that are interested in Lingual-SMoE are comparing it to the libraries listed below
Sorting:
- ACL'2025: SoftCoT: Soft Chain-of-Thought for Efficient Reasoning with LLMs. and preprint: SoftCoT++: Test-Time Scaling with Soft Chain-of…☆35Updated last month
- The this is the official implementation of "DAPE: Data-Adaptive Positional Encoding for Length Extrapolation"☆38Updated 9 months ago
- [NeurIPS'24 Oral] HydraLoRA: An Asymmetric LoRA Architecture for Efficient Fine-Tuning☆216Updated 7 months ago
- [ICML 2024] Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibrati…☆40Updated last year
- Official repository for paper "DeepCritic: Deliberate Critique with Large Language Models"☆32Updated 3 weeks ago
- ☆91Updated 2 months ago
- AdaMoLE: Adaptive Mixture of LoRA Experts☆34Updated 9 months ago
- ☆26Updated last year
- An Efficient LLM Fine-Tuning Factory Optimized for MoE PEFT☆106Updated 4 months ago
- Official PyTorch implementation for "Your Absorbing Discrete Diffusion Secretly Models the Conditional Distributions of Clean Data" (ICLR…☆55Updated last month
- Recent Advances on MLLM's Reasoning Ability☆24Updated 3 months ago
- [NeurIPS 2024] Code for the paper "Diffusion of Thoughts: Chain-of-Thought Reasoning in Diffusion Language Models"☆169Updated 4 months ago
- ☆147Updated 10 months ago
- AnchorAttention: Improved attention for LLMs long-context training☆211Updated 6 months ago
- paper list, tutorial, and nano code snippet for Diffusion Large Language Models.☆85Updated 3 weeks ago
- [NeurIPS 2024 Spotlight] EMR-Merging: Tuning-Free High-Performance Model Merging☆59Updated 4 months ago
- ☆174Updated 3 weeks ago
- State-of-the-art Parameter-Efficient MoE Fine-tuning Method☆169Updated 10 months ago
- Official PyTorch implementation of the paper "Accelerating Diffusion Large Language Models with SlowFast Sampling: The Three Golden Princ…☆23Updated last week
- V1: Toward Multimodal Reasoning by Designing Auxiliary Task☆34Updated 3 months ago
- Official implementation of "MMNeuron: Discovering Neuron-Level Domain-Specific Interpretation in Multimodal Large Language Model". Our co…☆19Updated 7 months ago
- [ICLR 2025] Dynamic Mixture of Experts: An Auto-Tuning Approach for Efficient Transformer Models☆118Updated last week
- [ICLR 2025] When Attention Sink Emerges in Language Models: An Empirical View (Spotlight)☆99Updated last week
- A Collection of Papers on Diffusion Language Models☆90Updated 2 weeks ago
- ☆137Updated last year
- [ICLR2025] DiffuGPT and DiffuLLaMA: Scaling Diffusion Language Models via Adaptation from Autoregressive Models☆242Updated last month
- 🚀 LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training☆86Updated 7 months ago
- ☆31Updated 6 months ago
- [ICML'25] Our study systematically investigates massive values in LLMs' attention mechanisms. First, we observe massive values are concen…☆74Updated 3 weeks ago
- [EMNLP 2023, Main Conference] Sparse Low-rank Adaptation of Pre-trained Language Models☆79Updated last year