intuit-ai-research / SPUQLinks
SPUQ: Perturbation-Based Uncertainty Quantification for Large Language Models
☆16Updated last year
Alternatives and similar repositories for SPUQ
Users that are interested in SPUQ are comparing it to the libraries listed below
Sorting:
- Benchmarking LLMs via Uncertainty Quantification☆246Updated last year
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"☆117Updated last year
- code repo for ICLR 2024 paper "Can LLMs Express Their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs"☆133Updated last year
- Official code for "Decoding-Time Language Model Alignment with Multiple Objectives".☆25Updated 11 months ago
- Lightweight Adapting for Black-Box Large Language Models☆23Updated last year
- [ICLR 2025] Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization☆31Updated 8 months ago
- ☆15Updated last year
- ICML 2024 - Official Repository for EXO: Towards Efficient Exact Optimization of Language Model Alignment☆57Updated last year
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆88Updated last year
- Code for paper: Aligning Large Language Models with Representation Editing: A Control Perspective☆34Updated 8 months ago
- ☆52Updated 6 months ago
- PaCE: Parsimonious Concept Engineering for Large Language Models (NeurIPS 2024)☆40Updated 11 months ago
- ☆30Updated last year
- This repository contains the code and data for the paper "SelfIE: Self-Interpretation of Large Language Model Embeddings" by Haozhe Chen,…☆51Updated 10 months ago
- [ACL 2024] Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models☆57Updated last year
- Is In-Context Learning Sufficient for Instruction Following in LLMs? [ICLR 2025]☆31Updated 8 months ago
- Principled Data Selection for Alignment: The Hidden Risks of Difficult Examples☆44Updated 2 months ago
- [NeurIPS 2023 Spotlight] Temperature Balancing, Layer-wise Weight Analysis, and Neural Network Training☆35Updated 6 months ago
- ☆38Updated last year
- This is the official repo for Towards Uncertainty-Aware Language Agent.☆28Updated last year
- ☆56Updated 2 years ago
- [ACL 25] SafeChain: Safety of Language Models with Long Chain-of-Thought Reasoning Capabilities☆19Updated 6 months ago
- ☆45Updated last year
- A Sober Look at Language Model Reasoning☆83Updated last week
- ☆34Updated last year
- Personalized Soups: Personalized Large Language Model Alignment via Post-hoc Parameter Merging☆108Updated last year
- This is an official implementation of the paper ``Building Math Agents with Multi-Turn Iterative Preference Learning'' with multi-turn DP…☆30Updated 10 months ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆80Updated 7 months ago
- Discriminative Constrained Optimization for Reinforcing Large Reasoning Models☆37Updated 3 weeks ago
- ☆97Updated last year