intuit-ai-research / SPUQLinks
SPUQ: Perturbation-Based Uncertainty Quantification for Large Language Models
☆16Updated last year
Alternatives and similar repositories for SPUQ
Users that are interested in SPUQ are comparing it to the libraries listed below
Sorting:
- Benchmarking LLMs via Uncertainty Quantification☆247Updated last year
- code repo for ICLR 2024 paper "Can LLMs Express Their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs"☆135Updated last year
- [ICLR 2025] Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization☆31Updated 9 months ago
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"☆118Updated last year
- ☆102Updated last year
- ICML 2024 - Official Repository for EXO: Towards Efficient Exact Optimization of Language Model Alignment☆57Updated last year
- Lightweight Adapting for Black-Box Large Language Models☆23Updated last year
- [ACL 2024] Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models☆59Updated last year
- Official code for "Decoding-Time Language Model Alignment with Multiple Objectives".☆27Updated last year
- ☆40Updated last year
- [NeurIPS 2023 Spotlight] Temperature Balancing, Layer-wise Weight Analysis, and Neural Network Training☆35Updated 6 months ago
- Official repository for Beyond Binary Rewards: Training LMs to Reason about Their Uncertainty☆38Updated 2 months ago
- Code for paper: Aligning Large Language Models with Representation Editing: A Control Perspective☆34Updated 9 months ago
- Is In-Context Learning Sufficient for Instruction Following in LLMs? [ICLR 2025]☆31Updated 9 months ago
- ☆30Updated last year
- Code for "Towards Revealing the Mystery behind Chain of Thought: a Theoretical Perspective"☆20Updated 2 years ago
- ☆52Updated 6 months ago
- ☆57Updated 2 years ago
- DataInf: Efficiently Estimating Data Influence in LoRA-tuned LLMs and Diffusion Models (ICLR 2024)☆76Updated last year
- Principled Data Selection for Alignment: The Hidden Risks of Difficult Examples☆44Updated 3 months ago
- ☆46Updated last year
- ☆41Updated last year
- PaCE: Parsimonious Concept Engineering for Large Language Models (NeurIPS 2024)☆40Updated 11 months ago
- Reasoning Activation in LLMs via Small Model Transfer (NeurIPS 2025)☆19Updated 2 weeks ago
- Directional Preference Alignment☆57Updated last year
- This is an official implementation of the paper ``Building Math Agents with Multi-Turn Iterative Preference Learning'' with multi-turn DP…☆30Updated 10 months ago
- ☆244Updated 5 months ago
- Stanford NLP Python library for benchmarking the utility of LLM interpretability methods☆136Updated 4 months ago
- [ACL'24] Beyond One-Preference-Fits-All Alignment: Multi-Objective Direct Preference Optimization☆91Updated last year
- [ACL 25] SafeChain: Safety of Language Models with Long Chain-of-Thought Reasoning Capabilities☆23Updated 7 months ago