yongchao98 / PROMST
Automatic prompt optimization framework for multi-step agent tasks.
☆27Updated 3 months ago
Alternatives and similar repositories for PROMST:
Users that are interested in PROMST are comparing it to the libraries listed below
- ☆48Updated 11 months ago
- The source code and dataset mentioned in the paper Seal-Tools: Self-Instruct Tool Learning Dataset for Agent Tuning and Detailed Benchmar…☆42Updated 3 months ago
- Hammer: Robust Function-Calling for On-Device Language Models via Function Masking☆58Updated this week
- [ICLR'24 spotlight] Tool-Augmented Reward Modeling☆44Updated last month
- We aim to provide the best references to search, select, and synthesize high-quality and large-quantity data for post-training your LLMs.☆49Updated 4 months ago
- ☆42Updated 2 months ago
- A scalable automated alignment method for large language models. Resources for "Aligning Large Language Models via Self-Steering Optimiza…☆14Updated 3 months ago
- Official implementation of the paper "From Complex to Simple: Enhancing Multi-Constraint Complex Instruction Following Ability of Large L…☆44Updated 7 months ago
- SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning. COLM 2024 Accepted Paper☆29Updated 8 months ago
- ☆13Updated 11 months ago
- [NeurIPS 2024] Train LLMs with diverse system messages reflecting individualized preferences to generalize to unseen system messages☆42Updated 2 months ago
- ☆15Updated 6 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆72Updated 8 months ago
- Codebase for Instruction Following without Instruction Tuning☆33Updated 4 months ago
- Implementations of online merging optimizers proposed by Online Merging Optimizers for Boosting Rewards and Mitigating Tax in Alignment☆73Updated 8 months ago
- Source code of "Reasons to Reject? Aligning Language Models with Judgments"☆58Updated 11 months ago
- Towards Systematic Measurement for Long Text Quality☆31Updated 5 months ago
- ☆23Updated last month
- This the implementation of LeCo☆30Updated last month
- Aligning with Human Judgement: The Role of Pairwise Preference in Large Language Model Evaluators (Liu et al.; COLM 2024)☆43Updated last month
- This repository contains the joint use of CPO and SimPO method for better reference-free preference learning methods.☆49Updated 6 months ago
- The code of arxiv paper: "CoT-based Synthesizer: Enhancing LLM Performance through Answer Synthesis"☆21Updated last month
- The official implementation of paper "Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language Models as Agen…☆23Updated 11 months ago
- Code for preprint "Metadata Conditioning Accelerates Language Model Pre-training (MeCo)"☆34Updated last month
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆46Updated last year
- Reformatted Alignment☆114Updated 4 months ago
- ☆45Updated 8 months ago