DeepSoftwareAnalytics / swe-factoryLinks
SWE-Factory: Your Automated Factory for Issue Resolution Training Data and Evaluation Benchmarks
β100Updated last month
Alternatives and similar repositories for swe-factory
Users that are interested in swe-factory are comparing it to the libraries listed below
Sorting:
- Official codebase for "GenPRM: Scaling Test-Time Compute of Process Reward Models via Generative Reasoning".β85Updated 4 months ago
- π§Tool-Star: Empowering LLM-brained Multi-Tool Reasoner via Reinforcement Learningβ282Updated last week
- RM-R1: Unleashing the Reasoning Potential of Reward Modelsβ146Updated 4 months ago
- End-to-End Reinforcement Learning for Multi-Turn Tool-Integrated Reasoningβ314Updated last month
- β142Updated last week
- Code, benchmark and environment for "ScienceBoard: Evaluating Multimodal Autonomous Agents in Realistic Scientific Workflows"β117Updated 2 months ago
- [NeurIPS 2025] The official repo of SynLogic: Synthesizing Verifiable Reasoning Data at Scale for Learning Logical Reasoning and Beyondβ176Updated 3 months ago
- Official Implementation of ARPO: End-to-End Policy Optimization for GUI Agents with Experience Replayβ133Updated 5 months ago
- β157Updated 3 weeks ago
- Revisiting Mid-training in the Era of Reinforcement Learning Scalingβ179Updated 3 months ago
- The official repo for "AceCoder: Acing Coder RL via Automated Test-Case Synthesis" [ACL25]β92Updated 6 months ago
- [ACL' 25] The official code repository for PRMBench: A Fine-grained and Challenging Benchmark for Process-Level Reward Models.β83Updated 8 months ago
- instruction-following benchmark for large reasoning modelsβ45Updated 2 months ago
- Test-time preferenece optimization (ICML 2025).β168Updated 5 months ago
- β133Updated last month
- A unified suite for generating elite reasoning problems and training high-performance LLMs, including pioneering attention-free architectβ¦β123Updated last week
- [ICML 2025] Teaching Language Models to Critique via Reinforcement Learningβ114Updated 5 months ago
- Official repository for paper: O1-Pruner: Length-Harmonizing Fine-Tuning for O1-Like Reasoning Pruningβ96Updated 8 months ago
- A comrephensive collection of learning from rewards in the post-training and test-time scaling of LLMs, with a focus on both reward modelβ¦β57Updated 4 months ago
- REverse-Engineered Reasoning for Open-Ended Generationβ77Updated last month
- [ICML 2025] M-STAR (Multimodal Self-Evolving TrAining for Reasoning) Project. Diving into Self-Evolving Training for Multimodal Reasoningβ69Updated 3 months ago
- β335Updated 3 months ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learningβ259Updated 5 months ago
- Towards a Unified View of Large Language Model Post-Trainingβ170Updated last month
- Laser: Learn to Reason Efficiently with Adaptive Length-based Reward Shapingβ55Updated 5 months ago
- SWE-Swiss: A Multi-Task Fine-Tuning and RL Recipe for High-Performance Issue Resolutionβ93Updated last month
- General Reasoner: Advancing LLM Reasoning Across All Domains [NeurIPS25]β192Updated this week
- β63Updated 4 months ago
- β46Updated 6 months ago
- [NeurIPS 2025 D&B] π SWE-bench Goes Live!β129Updated this week