MiaoXiong2320 / llm-uncertainty
code repo for ICLR 2024 paper "Can LLMs Express Their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs"
☆68Updated 7 months ago
Related projects ⓘ
Alternatives and complementary repositories for llm-uncertainty
- [ACL 2024] Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models☆35Updated 2 months ago
- Lightweight Adapting for Black-Box Large Language Models☆18Updated 8 months ago
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆45Updated 7 months ago
- ☆33Updated last year
- [NeurIPS'23] Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors☆69Updated 8 months ago
- ☆24Updated 5 months ago
- Reproduction of "RLCD Reinforcement Learning from Contrast Distillation for Language Model Alignment☆64Updated last year
- [ICLR 2024] Unveiling the Pitfalls of Knowledge Editing for Large Language Models☆22Updated 4 months ago
- This is the official repo for Towards Uncertainty-Aware Language Agent.☆22Updated 2 months ago
- ☆41Updated last year
- [NeurIPS 2024] Knowledge Circuits in Pretrained Transformers☆66Updated 3 weeks ago
- ☆38Updated last year
- ☆19Updated last month
- Implementation of PaCE: Parsimonious Concept Engineering for Large Language Models (NeurIPS 2024)☆26Updated this week
- ☆75Updated 4 months ago
- [ACL 2024] Code and data for "Machine Unlearning of Pre-trained Large Language Models"☆46Updated last month
- Offical code of the paper Large Language Models Are Implicitly Topic Models: Explaining and Finding Good Demonstrations for In-Context Le…☆68Updated 7 months ago
- ☆25Updated 9 months ago
- This paper list focuses on the theoretical and empirical analysis of language models, especially large language models (LLMs). The papers…☆51Updated this week
- ☆44Updated 10 months ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆51Updated this week
- [ATTRIB @ NeurIPS 2024] When Attention Sink Emerges in Language Models: An Empirical View☆27Updated 3 weeks ago
- ☆26Updated 6 months ago
- The Paper List on Data Contamination for Large Language Models Evaluation.☆73Updated this week
- Official code for "Decoding-Time Language Model Alignment with Multiple Objectives".☆12Updated last week
- ☆136Updated 4 months ago
- ☆78Updated last year
- DataInf: Efficiently Estimating Data Influence in LoRA-tuned LLMs and Diffusion Models (ICLR 2024)☆51Updated last month
- AdaICL: Which Examples to Annotate of In-Context Learning? Towards Effective and Efficient Selection☆16Updated last year
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆83Updated 5 months ago