Tongyi-Zhiwen / QwenLong-L1Links
☆287Updated 2 months ago
Alternatives and similar repositories for QwenLong-L1
Users that are interested in QwenLong-L1 are comparing it to the libraries listed below
Sorting:
- ☆77Updated 3 months ago
- ☆157Updated 3 months ago
- [ICML 2025] |TokenSwift: Lossless Acceleration of Ultra Long Sequence Generation☆110Updated 2 months ago
- ☆90Updated 2 months ago
- [ICML 2025] Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale☆255Updated 3 weeks ago
- ☆320Updated 10 months ago
- Deep Reasoning Translation (DRT) Project☆227Updated 2 months ago
- Efficient Agent Training for Computer Use☆120Updated last month
- A highly capable 2.4B lightweight LLM using only 1T pre-training data with all details.☆197Updated last week
- Parallel Scaling Law for Language Model — Beyond Parameter and Inference Time Scaling☆417Updated 2 months ago
- DeepResearch Bench: A Comprehensive Benchmark for Deep Research Agents☆233Updated last week
- Implementation for OAgents: An Empirical Study of Building Effective Agents☆98Updated this week
- ☆94Updated 7 months ago
- Repo for "MaskSearch: A Universal Pre-Training Framework to Enhance Agentic Search Capability"☆136Updated 2 months ago
- The official repo of SynLogic: Synthesizing Verifiable Reasoning Data at Scale for Learning Logical Reasoning and Beyond☆160Updated 3 weeks ago
- Towards Economical Inference: Enabling DeepSeek's Multi-Head Latent Attention in Any Transformer-based LLMs☆181Updated last month
- Mixture-of-Experts (MoE) Language Model☆189Updated 10 months ago
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆136Updated last year
- Ling is a MoE LLM provided and open-sourced by InclusionAI.☆181Updated 2 months ago
- Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models☆217Updated last month
- ☆66Updated 2 months ago
- AutoCoA (Automatic generation of Chain-of-Action) is an agent model framework that enhances the multi-turn tool usage capability of reaso…☆121Updated 4 months ago
- MMSearch-R1 is an end-to-end RL framework that enables LMMs to perform on-demand, multi-turn search with real-world multimodal search too…☆268Updated 3 weeks ago
- The RedStone repository includes code for preparing extensive datasets used in training large language models.☆136Updated last month
- A MemAgent framework that can be extrapolated to 3.5M, along with a training framework for RL training of any agent workflow.☆506Updated last week
- A Large-Scale, Challenging, Decontaminated, and Verifiable Mathematical Dataset for Advancing Reasoning☆239Updated last month
- ☆84Updated last week
- Agent KB: Leveraging Cross-Domain Experience for Agentic Problem Solving☆98Updated 2 weeks ago
- Delta-CoMe can achieve near loss-less 1-bit compressin which has been accepted by NeurIPS 2024☆57Updated 8 months ago
- Code for "Critique Fine-Tuning: Learning to Critique is More Effective than Learning to Imitate" [COLM 2025]☆168Updated 3 weeks ago