kailas-v / human-ai-interactions
☆11Updated 2 years ago
Alternatives and similar repositories for human-ai-interactions:
Users that are interested in human-ai-interactions are comparing it to the libraries listed below
- CEBaB: Estimating the Causal Effects of Real-World Concepts on NLP Model Behavior☆12Updated 2 years ago
- ☆23Updated last year
- This repository contains some of the code used in the paper "Training Language Models with Langauge Feedback at Scale"☆27Updated last year
- Code for the ICLR 2021 Paper "In-N-Out: Pre-Training and Self-Training using Auxiliary Information for Out-of-Distribution Robustness"☆12Updated 3 years ago
- csl: PyTorch-based Constrained Learning☆12Updated 2 years ago
- ☆38Updated last year
- Code for Residual Energy-Based Models for Text Generation in PyTorch.☆23Updated 3 years ago
- [ICLR 2025] Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization☆23Updated 2 months ago
- Official repository for our paper, Transformers Learn Higher-Order Optimization Methods for In-Context Learning: A Study with Linear Mode…☆15Updated 4 months ago
- Experiments and code to generate the GINC small-scale in-context learning dataset from "An Explanation for In-context Learning as Implici…☆104Updated last year
- Extending Conformal Prediction to LLMs☆64Updated 9 months ago
- Post-processing for fair classification☆13Updated 2 months ago
- Conformal Language Modeling☆28Updated last year
- Code for the Population-Based Bandits Algorithm, presented at NeurIPS 2020.☆20Updated 3 years ago
- Beta Shapley: a Unified and Noise-reduced Data Valuation Framework for Machine Learning (AISTATS 2022 Oral)☆40Updated 2 years ago
- ☆28Updated last year
- In-context Example Selection with Influences☆15Updated last year
- Align your LM to express calibrated verbal statements of confidence in its long-form generations.☆22Updated 9 months ago
- Is In-Context Learning Sufficient for Instruction Following in LLMs? [ICLR 2025]☆29Updated 2 months ago
- ☆31Updated last year
- The code of paper "Toward Optimal LLM Alignments Using Two-Player Games".☆16Updated 9 months ago
- ☆16Updated last year
- ☆37Updated last year
- Fairer Preferences Elicit Improved Human-Aligned Large Language Model Judgments (Zhou et al., EMNLP 2024)☆13Updated 5 months ago
- A repo for RLHF training and BoN over LLMs, with support for reward model ensembles.☆41Updated 2 months ago
- ☆47Updated last year
- Group-conditional DRO to alleviate spurious correlations☆15Updated 3 years ago
- ☆60Updated 3 years ago
- Tools for robustness evaluation in interpretability methods☆10Updated 3 years ago
- Compositional Explanations of Neurons, NeurIPS 2020 https://arxiv.org/abs/2006.14032☆25Updated 3 years ago