ByteDance-Seed / Seed-Coder
Seed-Coder is a family of open-source code LLMs comprising base, instruct and reasoning models of 8B size, developed by ByteDance Seed.
☆183Updated this week
Alternatives and similar repositories for Seed-Coder
Users that are interested in Seed-Coder are comparing it to the libraries listed below
Sorting:
- Scaling Data for SWE-agents☆101Updated this week
- An Open Math Pre-trainng Dataset with 370B Tokens.☆80Updated last month
- Official PyTorch implementation for Hogwild! Inference: Parallel LLM Generation with a Concurrent Attention Cache☆99Updated 2 weeks ago
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆188Updated last week
- Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-based LLMs☆164Updated last week
- A Large-Scale, Challenging, Decontaminated, and Verifiable Mathematical Dataset for Advancing Reasoning☆180Updated last week
- ☆65Updated last month
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆172Updated 3 months ago
- Tina: Tiny Reasoning Models via LoRA☆192Updated 2 weeks ago
- Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"☆518Updated last month
- Deep Reasoning Translation via Reinforcement Learning (arXiv preprint 2025); DRT: Deep Reasoning Translation via Long Chain-of-Thought (a…☆219Updated 2 weeks ago
- Prompt-to-Leaderboard☆222Updated this week
- CursorCore: Assist Programming through Aligning Anything☆122Updated 2 months ago
- Computer Agent Arena: Test & compare AI agents in real desktop apps & web environments. Code/data coming soon!☆44Updated last month
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate"☆144Updated 3 weeks ago
- ☆76Updated last month
- ☆131Updated this week
- ☆79Updated 2 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆323Updated 5 months ago
- official repository for “Reinforcement Learning for Reasoning in Large Language Models with One Training Example”☆143Updated last week
- Simple extension on vLLM to help you speed up reasoning model without training.☆149Updated last week
- ☆54Updated last month
- ☆93Updated 3 months ago
- ☆86Updated this week
- Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆90Updated 2 months ago
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆174Updated last month
- GRadient-INformed MoE☆262Updated 7 months ago
- ☆84Updated last week
- Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models☆162Updated this week
- PyTorch building blocks for the OLMo ecosystem☆210Updated this week