Tongyi-Zhiwen / QwenLong-L1Links
☆131Updated this week
Alternatives and similar repositories for QwenLong-L1
Users that are interested in QwenLong-L1 are comparing it to the libraries listed below
Sorting:
- ☆80Updated 2 weeks ago
- ☆76Updated last month
- An Open Math Pre-trainng Dataset with 370B Tokens.☆87Updated last month
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆132Updated 11 months ago
- [ICML 2025] Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale☆248Updated 2 weeks ago
- Parallel Scaling Law for Language Model — Beyond Parameter and Inference Time Scaling☆345Updated 2 weeks ago
- ☆150Updated last month
- Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-based LLMs☆166Updated last week
- Deep Reasoning Translation via Reinforcement Learning (arXiv preprint 2025); DRT: Deep Reasoning Translation via Long Chain-of-Thought (a…☆222Updated this week
- Efficient Agent Training for Computer Use☆85Updated this week
- IKEA: Reinforced Internal-External Knowledge Synergistic Reasoning for Efficient Adaptive Search Agent☆54Updated 2 weeks ago
- [ICML 2025] |TokenSwift: Lossless Acceleration of Ultra Long Sequence Generation☆99Updated last week
- ☆94Updated 5 months ago
- Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper☆136Updated 10 months ago
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆179Updated 2 months ago
- A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.☆182Updated this week
- General Reasoner: Advancing LLM Reasoning Across All Domains☆117Updated this week
- A Comprehensive Survey on Long Context Language Modeling☆146Updated last week
- Reformatted Alignment☆113Updated 8 months ago
- Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models☆166Updated last week
- ☆93Updated 3 months ago
- Advancing Language Model Reasoning through Reinforcement Learning and Inference Scaling☆101Updated 4 months ago
- Ling is a MoE LLM provided and open-sourced by InclusionAI.☆156Updated 2 weeks ago
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate"☆151Updated last month
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.☆238Updated last month
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆201Updated 3 weeks ago
- ☆104Updated last year
- ☆100Updated last month
- Official implementation for "ScoreFlow: Mastering LLM Agent Workflows via Score-based Preference Optimization"☆74Updated last week
- ☆82Updated last year