HKUST-KnowComp / PrivaCI-BenchLinks
☆12Updated 3 months ago
Alternatives and similar repositories for PrivaCI-Bench
Users that are interested in PrivaCI-Bench are comparing it to the libraries listed below
Sorting:
- Can Knowledge Editing Really Correct Hallucinations? (ICLR 2025)☆24Updated 2 months ago
- code repo for ICLR 2024 paper "Can LLMs Express Their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs"☆125Updated last year
- ☆33Updated 9 months ago
- Code release for "Debating with More Persuasive LLMs Leads to More Truthful Answers"☆113Updated last year
- [ICLR'24 Spotlight] A language model (LM)-based emulation framework for identifying the risks of LM agents with tool use☆152Updated last year
- ☆28Updated last year
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆96Updated last year
- ☆43Updated 5 months ago
- Implementation of the MATRIX framework (ICML 2024)☆57Updated last year
- [EMNLP 2023] Poisoning Retrieval Corpora by Injecting Adversarial Passages https://arxiv.org/abs/2310.19156☆35Updated last year
- Using Explanations as a Tool for Advanced LLMs☆66Updated 10 months ago
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆61Updated 8 months ago
- ☆24Updated last year
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agents (EMNLP Findings 2024)☆82Updated 2 months ago
- ☆99Updated 3 months ago
- This repository contains the code and data for the paper "SelfIE: Self-Interpretation of Large Language Model Embeddings" by Haozhe Chen,…☆50Updated 7 months ago
- BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).☆151Updated last year
- ☆24Updated 9 months ago
- On Memorization of Large Language Models in Logical Reasoning☆70Updated 4 months ago
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆61Updated last year
- [ACL'24] Chain of Thought (CoT) is significant in improving the reasoning abilities of large language models (LLMs). However, the correla…☆46Updated 2 months ago
- Interpretable Contrastive Monte Carlo Tree Search Reasoning☆48Updated 8 months ago
- e☆39Updated 3 months ago
- ☆23Updated 2 months ago
- ☆44Updated 5 months ago
- [NAACL 2025] The official implementation of paper "Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language M…☆26Updated last year
- [NeurIPS 2024] Knowledge Circuits in Pretrained Transformers☆151Updated 5 months ago
- Dialogue Action Tokens: Steering Language Models in Goal-Directed Dialogue with a Multi-Turn Planner☆26Updated last year
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examples☆103Updated last week
- Awesome LLM Self-Consistency: a curated list of Self-consistency in Large Language Models☆105Updated 2 weeks ago