QwenLM / QwQLinks
QwQ is the reasoning model series developed by Qwen team, Alibaba Cloud.
☆518Updated 5 months ago
Alternatives and similar repositories for QwQ
Users that are interested in QwQ are comparing it to the libraries listed below
Sorting:
- This repository introduce a comprehensive paper list, datasets, methods and tools for memory research.☆270Updated 3 months ago
- Train your Agent model via our easy and efficient framework☆1,381Updated this week
- ☆854Updated this week
- Moxin is a family of fully open-source and reproducible LLMs☆611Updated 2 months ago
- ☆396Updated this week
- adds Sequence Parallelism into LLaMA-Factory☆552Updated this week
- The official implementation of Self-Play Preference Optimization (SPPO)☆577Updated 7 months ago
- verl-agent is an extension of veRL, designed for training LLM/VLM agents via RL. verl-agent is also the official code for paper "Group-in…☆824Updated this week
- UI-Venus is a native UI agent designed to perform precise GUI element grounding and effective navigation using only screenshots as input.☆430Updated last week
- ☆812Updated 2 months ago
- A scalable, end-to-end training pipeline for general-purpose agents☆357Updated 2 months ago
- Think Beyond Images☆452Updated last week
- [COLM'25] DeepRetrieval - 🔥 Training Search Agent with Retrieval Outcomes via Reinforcement Learning☆624Updated 2 months ago
- Scaling RL on advanced reasoning models☆580Updated 3 weeks ago
- ☆293Updated 3 months ago
- AN O1 REPLICATION FOR CODING☆334Updated 8 months ago
- Parallel Scaling Law for Language Model — Beyond Parameter and Inference Time Scaling☆435Updated 3 months ago
- Official codebase for "Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling".☆271Updated 6 months ago
- Kimi-VL: Mixture-of-Experts Vision-Language Model for Multimodal Reasoning, Long-Context Understanding, and Strong Agent Capabilities☆1,051Updated last month
- The official implementation of the ICML 2024 paper "MemoryLLM: Towards Self-Updatable Large Language Models" and "M+: Extending MemoryLLM…☆214Updated last month
- [NeurIPS 2024] BAdam: A Memory Efficient Full Parameter Optimization Method for Large Language Models☆267Updated 5 months ago
- minimal-cost for training 0.5B R1-Zero☆767Updated 3 months ago
- Seed-Coder is a family of lightweight open-source code LLMs comprising base, instruct and reasoning models, developed by ByteDance Seed.☆549Updated 3 months ago
- Build, manage, and scale your AI agents with ease.☆432Updated this week
- Speed Always Wins: A Survey on Efficient Architectures for Large Language Models☆308Updated last week
- ReasonFlux Series - A family of LLM post-training algorithms focusing on data selection, reinforcement learning, and inference scaling☆484Updated last month
- ✨ A synthetic dataset generation framework that produces diverse coding questions and verifiable solutions - all in one framwork☆259Updated last month
- [EMNLP'25] s3 - ⚡ Efficient & Effective Search Agent Training via RL for RAG (Verifier-Powered RLVR for Search)☆638Updated last month
- ☆740Updated this week
- Recipes to train the self-rewarding reasoning LLMs.☆226Updated 6 months ago