Tongyi-Zhiwen / QwenLong-L1Links
☆269Updated 3 weeks ago
Alternatives and similar repositories for QwenLong-L1
Users that are interested in QwenLong-L1 are comparing it to the libraries listed below
Sorting:
- ☆152Updated last month
- ☆77Updated 2 months ago
- ☆86Updated last month
- [ICML 2025] |TokenSwift: Lossless Acceleration of Ultra Long Sequence Generation☆103Updated last month
- Deep Reasoning Translation (DRT) Project☆225Updated 3 weeks ago
- [ICML 2025] Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale☆251Updated 2 weeks ago
- Parallel Scaling Law for Language Model — Beyond Parameter and Inference Time Scaling☆395Updated last month
- Efficient Agent Training for Computer Use☆104Updated 2 weeks ago
- IKEA: Reinforced Internal-External Knowledge Synergistic Reasoning for Efficient Adaptive Search Agent☆57Updated last month
- A Comprehensive Survey on Long Context Language Modeling☆151Updated 2 weeks ago
- Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-based LLMs☆176Updated this week
- Ling is a MoE LLM provided and open-sourced by InclusionAI.☆169Updated last month
- An Open Math Pre-trainng Dataset with 370B Tokens.☆89Updated 2 months ago
- ☆94Updated 6 months ago
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆133Updated last year
- Implementation of the LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens Paper☆137Updated 11 months ago
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.☆240Updated 2 months ago
- ☆297Updated 3 weeks ago
- General Reasoner: Advancing LLM Reasoning Across All Domains☆141Updated last week
- The RedStone repository includes code for preparing extensive datasets used in training large language models.☆135Updated last week
- DeepResearch Bench: A Comprehensive Benchmark for Deep Research Agents☆82Updated this week
- A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.☆190Updated 2 weeks ago
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate"☆157Updated 2 weeks ago
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆186Updated 3 months ago
- Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models☆169Updated last month
- AutoCoA (Automatic generation of Chain-of-Action) is an agent model framework that enhances the multi-turn tool usage capability of reaso…☆115Updated 3 months ago
- Repo for "VRAG-RL: Empower Vision-Perception-Based RAG for Visually Rich Information Understanding via Iterative Reasoning with Reinforce…☆236Updated last week
- The All-in-one Judge Models introduced by Opencompass☆93Updated 3 months ago
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆219Updated last month
- Reformatted Alignment☆113Updated 8 months ago