[NAACL'25 🏆 SAC Award] Official code for "Advancing MoE Efficiency: A Collaboration-Constrained Routing (C2R) Strategy for Better Expert Parallelism Design"
☆16Feb 4, 2025Updated last year
Alternatives and similar repositories for C2R-MoE
Users that are interested in C2R-MoE are comparing it to the libraries listed below
Sorting:
- [ICML‘25] Official code for paper "Occult: Optimizing Collaborative Communication across Experts for Accelerated Parallel MoE Training an…☆13Apr 17, 2025Updated 11 months ago
- Official code for the paper "HEXA-MoE: Efficient and Heterogeneous-Aware MoE Acceleration with Zero Computation Redundancy"☆15Mar 6, 2025Updated last year
- Kinetics: Rethinking Test-Time Scaling Laws☆86Jul 11, 2025Updated 8 months ago
- Efficient Expert Pruning for Sparse Mixture-of-Experts Language Models: Enhancing Performance and Reducing Inference Costs☆23Nov 11, 2025Updated 4 months ago
- ☆19Nov 5, 2024Updated last year
- [ICML 2024] Code for the paper "MoE-RBench: Towards Building Reliable Language Models with Sparse Mixture-of-Experts"☆10Jul 1, 2024Updated last year
- Low-Rank Llama Custom Training☆23Mar 27, 2024Updated last year
- [CoLM'25] The official implementation of the paper <MoA: Mixture of Sparse Attention for Automatic Large Language Model Compression>☆157Jan 14, 2026Updated 2 months ago
- a website for accessing many models through api(deepseek、Qwen、Hunyuan etc.)☆16Jul 12, 2025Updated 8 months ago
- [ICLR‘24 Spotlight] Code for the paper "Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing Policy"☆104Jun 20, 2025Updated 9 months ago
- [arxiv: 2512.19673] Bottom-up Policy Optimization: Your Language Model Policy Secretly Contains Internal Policies☆60Feb 6, 2026Updated last month
- ☆25Apr 10, 2025Updated 11 months ago
- Official code for the paper "Examining Post-Training Quantization for Mixture-of-Experts: A Benchmark"☆30Jun 30, 2025Updated 8 months ago
- [ICLR 2025] MoE++: Accelerating Mixture-of-Experts Methods with Zero-Computation Experts☆264Oct 16, 2024Updated last year
- USTC 随机过程资料☆26Jul 17, 2022Updated 3 years ago
- The official repository for the experiments included in the paper titled "Patch-level Routing in Mixture-of-Experts is Provably Sample-ef…☆14Feb 12, 2026Updated last month
- SJTU 中文简约 LaTeX 报告模板☆10Jun 7, 2021Updated 4 years ago
- 2022 秋季学期清华大学电子系数据与算法课程 OJ 参考解答☆10Jun 18, 2023Updated 2 years ago
- ☆37Jan 13, 2026Updated 2 months ago
- AdaSkip: Adaptive Sublayer Skipping for Accelerating Long-Context LLM Inference☆20Jan 24, 2025Updated last year
- This repo implements an interface to GTAV for SCENIC language.☆11Dec 7, 2019Updated 6 years ago
- Inverse Scaling in Test-Time Compute☆25Dec 3, 2025Updated 3 months ago
- [ICLR'25] ARB-LLM: Alternating Refined Binarizations for Large Language Models☆28Aug 5, 2025Updated 7 months ago
- [DAC 2024] EDGE-LLM: Enabling Efficient Large Language Model Adaptation on Edge Devices via Layerwise Unified Compression and Adaptive La…☆84Jun 30, 2024Updated last year
- Transformers components but in Triton☆34May 9, 2025Updated 10 months ago
- ✂️ EyeLipCropper is a Python tool to crop eyes and mouth ROIs of the given video.☆14Nov 28, 2021Updated 4 years ago
- Ongoing research training transformer models at scale☆18Updated this week
- [CVPRW2024, Official Code] for paper "Exploring AIGC Video Quality: A Focus on Visual Harmony, Video-Text Consistency and Domain Distribu…☆14Jun 14, 2024Updated last year
- HISIM introduces a suite of analytical models at the system level to speed up performance prediction for AI models, covering logic-on-log…☆64Mar 17, 2025Updated last year
- ☆22Jun 10, 2025Updated 9 months ago
- Currently, there are many DeepSeek API providers on the market. Use DeepSeek Api Test to test which API performs the best☆19Feb 13, 2025Updated last year
- Official implementation of "Reasoning by Superposition: A Theoretical Perspective on Chain of Continuous Thought" (NeurIPS 2025)☆38Oct 8, 2025Updated 5 months ago
- ☆17Aug 13, 2024Updated last year
- Implementation for the paper: CMoE: Fast Carving of Mixture-of-Experts for Efficient LLM Inference☆35Mar 6, 2025Updated last year
- ☆39Sep 13, 2025Updated 6 months ago
- Official implementation of paper "VMoBA: Mixture-of-Block Attention for Video Diffusion Models"☆62Jul 1, 2025Updated 8 months ago
- 2022龙芯杯个人赛三等奖作品☆14Oct 11, 2023Updated 2 years ago
- NSCSCC “龙芯杯” 2024 个人赛 LoongArch 赛道三等奖☆14Aug 17, 2024Updated last year
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆139Jun 12, 2024Updated last year