gao-g / prelude
Code for the paper "Aligning LLM Agents by Learning Latent Preference from User Edits".
☆27Updated last month
Related projects ⓘ
Alternatives and complementary repositories for prelude
- [ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"☆49Updated 8 months ago
- Personalized Soups: Personalized Large Language Model Alignment via Post-hoc Parameter Merging☆98Updated last year
- ☆46Updated 10 months ago
- ☆89Updated 11 months ago
- Repository for NPHardEval, a quantified-dynamic benchmark of LLMs☆48Updated 7 months ago
- Generating diverse counterfactual data for Natural Language Understanding tasks using Large Language Models (LLMs). The generator support…☆35Updated last year
- ☆25Updated 4 months ago
- Repository for paper Tools Are Instrumental for Language Agents in Complex Environments☆33Updated last month
- Trial and Error: Exploration-Based Trajectory Optimization of LLM Agents (ACL 2024 Main Conference)☆99Updated 3 weeks ago
- ☆29Updated 7 months ago
- GenRM-CoT: Data release for verification rationales☆24Updated last month
- Evaluate the Quality of Critique☆35Updated 5 months ago
- Code for the paper <SelfCheck: Using LLMs to Zero-Shot Check Their Own Step-by-Step Reasoning>☆45Updated last year
- Supporting code for ReCEval paper☆26Updated 2 months ago
- ☆28Updated last year
- Accompanying code for "Boosted Prompt Ensembles for Large Language Models"☆28Updated last year
- Directional Preference Alignment☆51Updated 2 months ago
- Repo for: When to Make Exceptions: Exploring Language Models as Accounts of Human Moral Judgment☆38Updated last year
- BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval☆57Updated last month
- Offical code of the paper Large Language Models Are Implicitly Topic Models: Explaining and Finding Good Demonstrations for In-Context Le…☆68Updated 8 months ago
- This repository contains data, code and models for contextual noncompliance.☆18Updated 4 months ago
- ☆103Updated 4 months ago
- Aligning with Human Judgement: The Role of Pairwise Preference in Large Language Model Evaluators (Liu et al.; arXiv preprint arXiv:2403.…☆37Updated 4 months ago
- ☆34Updated 3 months ago
- Interpretable text embeddings by asking LLMs yes/no questions (NeurIPS 2024)☆22Updated last week
- ☆41Updated last year
- Code/data for MARG (multi-agent review generation)☆33Updated last week
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"☆84Updated 8 months ago
- ☆73Updated 4 months ago
- Easy-to-Hard Generalization: Scalable Alignment Beyond Human Supervision☆97Updated 2 months ago