allenai / hybrid-preferencesLinks
Learning to route instances for Human vs AI Feedback (ACL 2025 Main)
☆23Updated 2 weeks ago
Alternatives and similar repositories for hybrid-preferences
Users that are interested in hybrid-preferences are comparing it to the libraries listed below
Sorting:
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆57Updated 9 months ago
- Aioli: A unified optimization framework for language model data mixing☆25Updated 4 months ago
- ReBase: Training Task Experts through Retrieval Based Distillation☆29Updated 4 months ago
- ☆49Updated 6 months ago
- ☆58Updated 2 weeks ago
- Measuring and Controlling Persona Drift in Language Model Dialogs☆17Updated last year
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆43Updated last year
- A framework for pitting LLMs against each other in an evolving library of games ⚔☆32Updated last month
- ☆57Updated 8 months ago
- Q-Probe: A Lightweight Approach to Reward Maximization for Language Models☆41Updated 11 months ago
- [ACL'24] Code and data of paper "When is Tree Search Useful for LLM Planning? It Depends on the Discriminator"☆54Updated last year
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆70Updated 11 months ago
- Learning to Retrieve by Trying - Source code for Grounding by Trying: LLMs with Reinforcement Learning-Enhanced Retrieval☆34Updated 7 months ago
- ☆42Updated 2 months ago
- ☆21Updated 3 months ago
- Middleware for LLMs: Tools Are Instrumental for Language Agents in Complex Environments (EMNLP'2024)☆36Updated 5 months ago
- The repository contains code for Adaptive Data Optimization☆24Updated 5 months ago
- Verifiers for LLM Reinforcement Learning☆55Updated last month
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- Codebase accompanying the Summary of a Haystack paper.☆78Updated 8 months ago
- Stanford NLP Python library for benchmarking the utility of LLM interpretability methods☆89Updated last week
- Code/data for MARG (multi-agent review generation)☆43Updated 6 months ago
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆55Updated 2 weeks ago
- [arXiv preprint] Official Repository for "Evaluating Language Models as Synthetic Data Generators"☆33Updated 5 months ago
- ☆11Updated 2 weeks ago
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆24Updated last year
- ☆24Updated last year
- This repository includes a benchmark and code for the paper "Evaluating LLMs at Detecting Errors in LLM Responses".☆29Updated 9 months ago
- CausalGym: Benchmarking causal interpretability methods on linguistic tasks☆43Updated 6 months ago
- ☆29Updated 2 weeks ago