lqtrung1998 / mwp_ReFT
☆504Updated 2 months ago
Alternatives and similar repositories for mwp_ReFT:
Users that are interested in mwp_ReFT are comparing it to the libraries listed below
- A series of technical report on Slow Thinking with LLM☆595Updated last week
- ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search (NeurIPS 2024)☆597Updated 2 months ago
- This is the repository that contains the source code for the Self-Evaluation Guided MCTS for online DPO.☆299Updated 7 months ago
- [NAACL'24] Self-data filtering of LLM instruction-tuning data using a novel perplexity-based difficulty score, without using any other mo…☆347Updated 6 months ago
- ☆325Updated last month
- ☆186Updated this week
- ☆559Updated 2 weeks ago
- The related works and background techniques about Openai o1☆217Updated 2 months ago
- R1-searcher: Incentivizing the Search Capability in LLMs via Reinforcement Learning☆376Updated this week
- Codes and Data for Scaling Relationship on Learning Mathematical Reasoning with Large Language Models☆251Updated 6 months ago
- [ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuning☆424Updated 5 months ago
- Implementation for "Step-DPO: Step-wise Preference Optimization for Long-chain Reasoning of LLMs"☆357Updated 2 months ago
- ☆910Updated 2 months ago
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]☆542Updated 3 months ago
- Large Reasoning Models☆799Updated 3 months ago
- A very simple GRPO implement for reproducing r1-like LLM thinking.☆782Updated this week
- OpenRFT: Adapting Reasoning Foundation Model for Domain-specific Tasks with Reinforcement Fine-Tuning☆124Updated 3 months ago
- ☆264Updated 8 months ago
- A visuailzation tool to make deep understaning and easier debugging for RLHF training.☆177Updated last month
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Reward☆851Updated last month
- [ACL 2024] A Survey of Chain of Thought Reasoning: Advances, Frontiers and Future☆430Updated 2 months ago
- Recipes to train reward model for RLHF.☆1,257Updated last month
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.☆212Updated this week
- InsTag: A Tool for Data Analysis in LLM Supervised Fine-tuning☆246Updated last year
- OpenR: An Open Source Framework for Advanced Reasoning with Large Language Models☆1,732Updated 2 months ago
- LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment☆311Updated 10 months ago
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data …☆660Updated last week
- ReSearch: Learning to Reason with Search for LLMs via Reinforcement Learning☆306Updated 3 weeks ago
- minimal-cost for training 0.5B R1-Zero☆668Updated 2 weeks ago
- LongBench v2 and LongBench (ACL 2024)☆819Updated 2 months ago