likenneth / dialogue_action_token
Dialogue Action Tokens: Steering Language Models in Goal-Directed Dialogue with a Multi-Turn Planner
☆21Updated 7 months ago
Alternatives and similar repositories for dialogue_action_token:
Users that are interested in dialogue_action_token are comparing it to the libraries listed below
- official implementation of paper "Process Reward Model with Q-value Rankings"☆48Updated 2 weeks ago
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆55Updated 2 months ago
- Evaluate the Quality of Critique☆35Updated 8 months ago
- Critique-out-Loud Reward Models☆52Updated 4 months ago
- [EMNLP'24 (Main)] DRPO(Dynamic Rewarding with Prompt Optimization) is a tuning-free approach for self-alignment. DRPO leverages a search-…☆20Updated 3 months ago
- [ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"☆54Updated 11 months ago
- Scalable Meta-Evaluation of LLMs as Evaluators☆43Updated last year
- GenRM-CoT: Data release for verification rationales☆47Updated 4 months ago
- ☆92Updated last month
- Directional Preference Alignment☆56Updated 4 months ago
- Trial and Error: Exploration-Based Trajectory Optimization of LLM Agents (ACL 2024 Main Conference)☆116Updated 3 months ago
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆44Updated last month
- Watch Every Step! LLM Agent Learning via Iterative Step-level Process Refinement (EMNLP 2024 Main Conference)☆52Updated 4 months ago
- Personalized Soups: Personalized Large Language Model Alignment via Post-hoc Parameter Merging☆99Updated last year
- AbstainQA, ACL 2024☆25Updated 4 months ago
- [ACL 2024] Self-Training with Direct Preference Optimization Improves Chain-of-Thought Reasoning☆39Updated 6 months ago
- Code and data used in the paper: "Training on Incorrect Synthetic Data via RL Scales LLM Math Reasoning Eight-Fold"☆29Updated 8 months ago
- ☆27Updated 11 months ago
- CodeUltraFeedback: aligning large language models to coding preferences☆68Updated 7 months ago
- Supporting code for ReCEval paper☆28Updated 5 months ago
- The official repository of "Improving Large Language Models via Fine-grained Reinforcement Learning with Minimum Editing Constraint"☆34Updated last year
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆115Updated 5 months ago
- ☆93Updated last year
- ☆33Updated 10 months ago
- Source code of "Reasons to Reject? Aligning Language Models with Judgments"☆58Updated 11 months ago
- NeurIPS 2024 tutorial on LLM Inference☆39Updated 2 months ago
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆46Updated last year
- Self-Supervised Alignment with Mutual Information☆16Updated 8 months ago
- [ACL 2024] Masked Thought: Simply Masking Partial Reasoning Steps Can Improve Mathematical Reasoning Learning of Language Models☆16Updated 7 months ago