Arstanley / Awesome-Trustworthy-RAGLinks
☆94Updated 5 months ago
Alternatives and similar repositories for Awesome-Trustworthy-RAG
Users that are interested in Awesome-Trustworthy-RAG are comparing it to the libraries listed below
Sorting:
- [EMNLP 2024] The official GitHub repo for the survey paper "Knowledge Conflicts for LLMs: A Survey"☆150Updated last year
- Paper list for the survey "Combating Misinformation in the Age of LLMs: Opportunities and Challenges" and the initiative "LLMs Meet Misin…☆104Updated last year
- LLM Unlearning☆178Updated 2 years ago
- ☆152Updated last month
- Toolkit for evaluating the trustworthiness of generative foundation models.☆123Updated 3 months ago
- The dataset and code for the ICLR 2024 paper "Can LLM-Generated Misinformation Be Detected?"☆79Updated last year
- The lastest paper about detection of LLM-generated text and code☆280Updated 5 months ago
- [NAACL2024] Attacks, Defenses and Evaluations for LLM Conversation Safety: A Survey☆107Updated last year
- LLM hallucination paper list☆327Updated last year
- A curated list of resources for activation engineering☆117Updated 2 months ago
- 【ACL 2024】 SALAD benchmark & MD-Judge☆167Updated 9 months ago
- The implement of paper:"ReDeEP: Detecting Hallucination in Retrieval-Augmented Generation via Mechanistic Interpretability"☆52Updated 6 months ago
- JAILJUDGE: A comprehensive evaluation benchmark which includes a wide range of risk scenarios with complex malicious prompts (e.g., synth…☆53Updated last year
- ☆175Updated last year
- ☆55Updated last year
- [EMNLP 2023] Explainable Claim Verification via Knowledge-Grounded Reasoning with Large Language Models☆25Updated 2 years ago
- ☆23Updated 4 months ago
- This is the repo for the survey of Bias and Fairness in IR with LLMs.☆59Updated 3 months ago
- [ACL 2024] Shifting Attention to Relevance: Towards the Predictive Uncertainty Quantification of Free-Form Large Language Models☆59Updated last year
- RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models. NeurIPS 2024☆85Updated last year
- R-Judge: Benchmarking Safety Risk Awareness for LLM Agents (EMNLP Findings 2024)☆93Updated 7 months ago
- ☆17Updated last year
- [ACL'25 Main] SelfElicit: Your Language Model Secretly Knows Where is the Relevant Evidence! | 让你的LLM更好地利用上下文文档:一个基于注意力的简单方案☆24Updated 9 months ago
- ☆38Updated 2 years ago
- Source code of our paper MIND, ACL 2024 Long Paper☆59Updated last month
- A survey on harmful fine-tuning attack for large language model☆224Updated 3 weeks ago
- ☆28Updated last year
- awesome SAE papers☆66Updated 6 months ago
- UP-TO-DATE LLM Watermark paper. 🔥🔥🔥☆367Updated last year
- [USENIX Security 2025] PoisonedRAG: Knowledge Corruption Attacks to Retrieval-Augmented Generation of Large Language Models☆219Updated 3 weeks ago