VITA-Group / o1-planningLinks
On The Planning Abilities of OpenAI's o1 Models: Feasibility, Optimality, and Generalizability
β40Updated 3 months ago
Alternatives and similar repositories for o1-planning
Users that are interested in o1-planning are comparing it to the libraries listed below
Sorting:
- β16Updated last year
- [NeurIPS-2024] π Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623β86Updated last year
- β19Updated 7 months ago
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"β48Updated last year
- Exploration of automated dataset selection approaches at large scales.β47Updated 7 months ago
- β29Updated last year
- The official repository for SkyLadder: Better and Faster Pretraining via Context Window Schedulingβ35Updated last month
- β49Updated 8 months ago
- [NeurIPS 2025 Spotlight] ReasonFlux-Coder: Open-Source LLM Coders with Co-Evolving Reinforcement Learningβ122Updated 3 weeks ago
- [NeurIPS 2024] OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AIβ105Updated 7 months ago
- implementation of dualformerβ20Updated 7 months ago
- Code for Paper: Harnessing Webpage Uis For Text Rich Visual Understandingβ52Updated 9 months ago
- The code implementation of MAGDi: Structured Distillation of Multi-Agent Interaction Graphs Improves Reasoning in Smaller Language Modelsβ¦β37Updated last year
- β40Updated 3 months ago
- [ACL2025 Findings] Benchmarking Multihop Multimodal Internet Agentsβ46Updated 7 months ago
- Co-LLM: Learning to Decode Collaboratively with Multiple Language Modelsβ121Updated last year
- β27Updated 3 months ago
- [ACL 2025] Are Your LLMs Capable of Stable Reasoning?β30Updated 2 months ago
- Implementation of the model: "Reka Core, Flash, and Edge: A Series of Powerful Multimodal Language Models" in PyTorchβ28Updated this week
- Tree prompting: easy-to-use scikit-learn interface for improved prompting.β41Updated last year
- β65Updated last year
- [NAACL 2025 Oral] Multimodal Needle in a Haystack (MMNeedle): Benchmarking Long-Context Capability of Multimodal Large Language Modelsβ49Updated 5 months ago
- official repo for the paper "Learning From Mistakes Makes LLM Better Reasoner"β58Updated last year
- JudgeLRM: Large Reasoning Models as a Judgeβ39Updated 3 weeks ago
- Extensive Self-Contrast Enables Feedback-Free Language Model Alignmentβ20Updated last year
- Codebase for Instruction Following without Instruction Tuningβ35Updated last year
- Code for paper: Long cOntext aliGnment via efficient preference Optimizationβ13Updated 7 months ago
- [ICLR 2025] LongPO: Long Context Self-Evolution of Large Language Models through Short-to-Long Preference Optimizationβ40Updated 7 months ago
- [NeurIPS 2024] A comprehensive benchmark for evaluating critique ability of LLMsβ46Updated 10 months ago
- This is the oficial repository for "Safer-Instruct: Aligning Language Models with Automated Preference Data"β17Updated last year