ZihanWang314 / CoE
Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models
☆162Updated this week
Alternatives and similar repositories for CoE
Users that are interested in CoE are comparing it to the libraries listed below
Sorting:
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learning☆198Updated last week
- Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-based LLMs☆164Updated last week
- Efficient triton implementation of Native Sparse Attention.☆144Updated last month
- ☆76Updated last month
- Research Code for preprint "Optimizing Test-Time Compute via Meta Reinforcement Finetuning".☆93Updated 2 months ago
- Block Transformer: Global-to-Local Language Modeling for Fast Inference (NeurIPS 2024)☆154Updated last month
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate"☆144Updated 3 weeks ago
- Simple extension on vLLM to help you speed up reasoning model without training.☆149Updated last week
- ☆82Updated this week
- Repo for "Z1: Efficient Test-time Scaling with Code"☆58Updated last month
- An Open Math Pre-trainng Dataset with 370B Tokens.☆80Updated last month
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆132Updated 11 months ago
- TokenSkip: Controllable Chain-of-Thought Compression in LLMs☆138Updated 2 months ago
- A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.☆177Updated last month
- ☆194Updated 2 months ago
- Unofficial implementation for the paper "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆158Updated 10 months ago
- ☆78Updated 3 months ago
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆188Updated last week
- SkyRL-v0: Train Real-World Long-Horizon Agents via Reinforcement Learning☆180Updated this week
- CPPO: Accelerating the Training of Group Relative Policy Optimization-Based Reasoning Models☆125Updated last week
- ☆46Updated 2 months ago
- A Comprehensive Survey on Long Context Language Modeling☆139Updated last month
- official repository for “Reinforcement Learning for Reasoning in Large Language Models with One Training Example”☆143Updated last week
- SepLLM: Accelerate Large Language Models by Compressing One Segment into One Separator☆71Updated 4 months ago
- ☆153Updated last month
- Implementation of 🥥 Coconut, Chain of Continuous Thought, in Pytorch☆166Updated 4 months ago
- Code for "Your Mixture-of-Experts LLM Is Secretly an Embedding Model For Free"☆68Updated 6 months ago
- 🔥 A minimal training framework for scaling FLA models☆128Updated this week
- Implementation of the paper: "Mixture-of-Depths: Dynamically allocating compute in transformer-based language models"☆91Updated this week
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆323Updated 5 months ago