deepseek-ai / DeepSeek-MoE
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language Models
☆1,409Updated last year
Alternatives and similar repositories for DeepSeek-MoE:
Users that are interested in DeepSeek-MoE are comparing it to the libraries listed below
- Scalable RL solution for advanced reasoning of language models☆1,262Updated this week
- Expert Specialized Fine-Tuning☆529Updated 4 months ago
- Large Reasoning Models☆801Updated 2 months ago
- OLMoE: Open Mixture-of-Experts Language Models☆613Updated 2 months ago
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Models☆1,449Updated 11 months ago
- O1 Replication Journey☆1,947Updated last month
- [NeurIPS'24 Spotlight, ICLR'25] To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which r…☆917Updated last week
- OpenR: An Open Source Framework for Advanced Reasoning with Large Language Models☆1,651Updated last month
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆821Updated this week
- [ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding☆1,194Updated 4 months ago
- ☆422Updated 6 months ago
- Official Implementation of EAGLE-1 (ICML'24) and EAGLE-2 (EMNLP'24)☆968Updated this week
- AllenAI's post-training codebase☆2,668Updated this week
- ☆917Updated 2 weeks ago
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.☆700Updated 4 months ago
- An Open Large Reasoning Model for Real-World Solutions☆1,444Updated 2 months ago
- FlashInfer: Kernel Library for LLM Serving☆2,111Updated this week
- verl: Volcano Engine Reinforcement Learning for LLMs☆3,513Updated this week
- ⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024)☆922Updated 2 months ago
- ☆890Updated 3 weeks ago
- Scalable toolkit for efficient model alignment☆722Updated this week
- DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models☆2,389Updated 10 months ago
- An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)☆4,895Updated this week
- ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search (NeurIPS 2024)☆575Updated last month
- Code for Quiet-STaR☆713Updated 6 months ago
- DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model☆4,769Updated 4 months ago
- A curated list of open-source projects related to DeepSeek Coder☆571Updated 10 months ago