lingchen0331 / UQ_ICLLinks
Uncertainty quantification for in-context learning of large language models
☆16Updated last year
Alternatives and similar repositories for UQ_ICL
Users that are interested in UQ_ICL are comparing it to the libraries listed below
Sorting:
- Offical code of the paper Large Language Models Are Implicitly Topic Models: Explaining and Finding Good Demonstrations for In-Context Le…☆75Updated last year
- ☆38Updated last year
- [ICLR 2025] Unintentional Unalignment: Likelihood Displacement in Direct Preference Optimization☆31Updated 8 months ago
- Is In-Context Learning Sufficient for Instruction Following in LLMs? [ICLR 2025]☆31Updated 8 months ago
- [NeurIPS 2024] Code and Data Repo for Paper "Embedding Trajectory for Out-of-Distribution Detection in Mathematical Reasoning"☆27Updated last year
- EMNLP 2024: Model Editing Harms General Abilities of Large Language Models: Regularization to the Rescue☆37Updated 4 months ago
- ☆50Updated last year
- Lightweight Adapting for Black-Box Large Language Models☆23Updated last year
- This is the official repo for Towards Uncertainty-Aware Language Agent.☆28Updated last year
- Align your LM to express calibrated verbal statements of confidence in its long-form generations.☆27Updated last year
- Models, data, and codes for the paper: MetaAligner: Towards Generalizable Multi-Objective Alignment of Language Models☆24Updated last year
- Official code for "Decoding-Time Language Model Alignment with Multiple Objectives".☆25Updated 11 months ago
- ☆100Updated last year
- Source code for the TMLR paper "Black-Box Prompt Learning for Pre-trained Language Models"☆56Updated 2 years ago
- [NeurIPS'23] Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors☆81Updated 9 months ago
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆61Updated last year
- [ACL'25] Mosaic-IT: Cost-Free Compositional Data Synthesis for Instruction Tuning☆20Updated 2 weeks ago
- A curated list of resources on Reinforcement Learning with Verifiable Rewards (RLVR) and the reasoning capability boundary of Large Langu…☆57Updated this week
- A Kernel-Based View of Language Model Fine-Tuning https://arxiv.org/abs/2210.05643☆78Updated 2 years ago
- Learning adapter weights from task descriptions☆19Updated last year
- Teaching Models to Express Their Uncertainty in Words☆39Updated 3 years ago
- ☆51Updated last year
- Implementation of Gradient Information Optimization (GIO) for effective and scalable training data selection☆14Updated 2 years ago
- ☆46Updated 2 years ago
- ☆45Updated last year
- Fairer Preferences Elicit Improved Human-Aligned Large Language Model Judgments (Zhou et al., EMNLP 2024)☆13Updated last year
- This is the official implementation of TAGCOS: Task-agnostic Gradient Clustered Coreset Selection for Instruction Tuning Data☆10Updated last year
- ☆40Updated last year
- [ICLR 2025] Code&Data for the paper "Super(ficial)-alignment: Strong Models May Deceive Weak Models in Weak-to-Strong Generalization"☆13Updated last year
- source code for NeurIPS'24 paper "HaloScope: Harnessing Unlabeled LLM Generations for Hallucination Detection"☆58Updated 6 months ago