baixianghuang / HalluEditBenchLinks
Can Knowledge Editing Really Correct Hallucinations? (ICLR 2025)
☆27Updated 3 months ago
Alternatives and similar repositories for HalluEditBench
Users that are interested in HalluEditBench are comparing it to the libraries listed below
Sorting:
- [ACL 2024] Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models☆59Updated last year
- In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation (ICML 2024)☆62Updated last year
- [ACL'24] Chain of Thought (CoT) is significant in improving the reasoning abilities of large language models (LLMs). However, the correla…☆46Updated 6 months ago
- ☆30Updated last year
- [ACL 2024] Code and data for "Machine Unlearning of Pre-trained Large Language Models"☆63Updated last year
- code repo for ICLR 2024 paper "Can LLMs Express Their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs"☆137Updated last year
- ☆25Updated 7 months ago
- [EMNLP 2024] The official GitHub repo for the survey paper "Knowledge Conflicts for LLMs: A Survey"☆148Updated last year
- [NeurIPS 2024] Knowledge Circuits in Pretrained Transformers☆159Updated last week
- [NeurIPS 2024] The official implementation of paper: Chain of Preference Optimization: Improving Chain-of-Thought Reasoning in LLMs.☆132Updated 8 months ago
- [NAACL'25 Oral] Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering☆67Updated 11 months ago
- Reasoning Activation in LLMs via Small Model Transfer (NeurIPS 2025)☆20Updated last month
- A Sober Look at Language Model Reasoning☆87Updated last week
- RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models. NeurIPS 2024☆86Updated last year
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆99Updated last year
- Search, Verify and Feedback: Towards Next Generation Post-training Paradigm of Foundation Models via Verifier Engineering☆62Updated 11 months ago
- ☆41Updated 2 years ago
- [NeurIPS 2024] Code and Data Repo for Paper "Embedding Trajectory for Out-of-Distribution Detection in Mathematical Reasoning"☆27Updated last year
- LLM Unlearning☆177Updated 2 years ago
- Toolkit for evaluating the trustworthiness of generative foundation models.☆123Updated 3 months ago
- Using Explanations as a Tool for Advanced LLMs☆69Updated last year
- [NeurIPS 2024] Uncertainty of Thoughts: Uncertainty-Aware Planning Enhances Information Seeking in Large Language Models☆103Updated last year
- [COLM'25] Missing Premise exacerbates Overthinking: Are Reasoning Models losing Critical Thinking Skill?☆35Updated 5 months ago
- ☆131Updated 8 months ago
- [EMNLP 2025 Main] ConceptVectors Benchmark and Code for the paper "Intrinsic Evaluation of Unlearning Using Parametric Knowledge Traces"☆38Updated 3 months ago
- ☆10Updated 7 months ago
- code for EMNLP 2024 paper: Neuron-Level Knowledge Attribution in Large Language Models☆47Updated last year
- [ACL2025 Best Paper] Language Models Resist Alignment☆36Updated 5 months ago
- Accepted LLM Papers in NeurIPS 2024☆37Updated last year
- ☆35Updated last year