shuoli90 / Rank-Calibration
This is the repo for constructing a comprehensive and rigorous evaluation framework for LLM calibration.
☆12Updated last year
Alternatives and similar repositories for Rank-Calibration:
Users that are interested in Rank-Calibration are comparing it to the libraries listed below
- Conformal Language Modeling☆28Updated last year
- ☆28Updated last month
- PaCE: Parsimonious Concept Engineering for Large Language Models (NeurIPS 2024)☆35Updated 5 months ago
- ☆42Updated last year
- In-context Example Selection with Influences☆15Updated last year
- ☆87Updated 9 months ago
- ☆50Updated 2 weeks ago
- [NeurIPS'23] Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors☆75Updated 4 months ago
- ☆28Updated last year
- ☆49Updated last year
- AbstainQA, ACL 2024☆25Updated 6 months ago
- ConceptVectors Benchmark and Code for the paper "Intrinsic Evaluation of Unlearning Using Parametric Knowledge Traces"☆35Updated 2 months ago
- Augmenting Statistical Models with Natural Language Parameters☆26Updated 7 months ago
- Is In-Context Learning Sufficient for Instruction Following in LLMs? [ICLR 2025]☆29Updated 3 months ago
- ☆46Updated last month
- LoFiT: Localized Fine-tuning on LLM Representations☆37Updated 3 months ago
- ☆40Updated last year
- ☆13Updated last year
- ☆82Updated 8 months ago
- Lightweight Adapting for Black-Box Large Language Models☆22Updated last year
- [NAACL'25 Oral] Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering☆53Updated 5 months ago
- ☆42Updated 2 months ago
- Repo for: When to Make Exceptions: Exploring Language Models as Accounts of Human Moral Judgment☆38Updated last year
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆72Updated last month
- Stanford NLP Python library for benchmarking the utility of LLM interpretability methods☆70Updated 3 weeks ago
- [ICLR 2025] Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization☆24Updated 3 months ago
- ☆27Updated last month
- Can Knowledge Editing Really Correct Hallucinations? (ICLR 2025)☆12Updated 2 months ago
- Unofficial implementation of Conformal Language Modeling by Quach et al☆28Updated last year
- Official repository for our paper, Transformers Learn Higher-Order Optimization Methods for In-Context Learning: A Study with Linear Mode…☆16Updated 5 months ago