ydyjya / Awesome-LLM-SafetyLinks
A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provide researchers, practitioners, and enthusiasts with insights into the safety implications, challenges, and advancements surrounding these powerful models.
☆1,449Updated 2 weeks ago
Alternatives and similar repositories for Awesome-LLM-Safety
Users that are interested in Awesome-LLM-Safety are comparing it to the libraries listed below
Sorting:
- A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).☆1,502Updated this week
- Papers and resources related to the security and privacy of LLMs 🤖☆511Updated 2 weeks ago
- A curation of awesome tools, documents and projects about LLM Security.☆1,259Updated 2 months ago
- 😎 up-to-date & curated list of awesome Attacks on Large-Vision-Language-Models papers, methods & resources.☆330Updated this week
- An Awesome Collection for LLM Survey☆366Updated last month
- awesome papers in LLM interpretability☆495Updated last week
- Reading list of hallucination in LLMs. Check out our new survey paper: "Siren’s Song in the AI Ocean: A Survey on Hallucination in Large …☆1,024Updated 7 months ago
- Must-read Papers on Knowledge Editing for Large Language Models.☆1,102Updated last week
- Latest Advances on System-2 Reasoning☆1,128Updated 3 weeks ago
- [ACL 2024] A Survey of Chain of Thought Reasoning: Advances, Frontiers and Future☆451Updated 5 months ago
- Safety at Scale: A Comprehensive Survey of Large Model Safety☆173Updated 4 months ago
- Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Models☆762Updated last month
- A resource repository for machine unlearning in large language models☆426Updated last month
- Latest Advances on Long Chain-of-Thought Reasoning☆390Updated last month
- Daily updated LLM papers. 每日更新 LLM 相关的论文,欢迎订阅 👏 喜欢的话动动你的小手 🌟 一个☆1,134Updated 10 months ago
- [ICLR 2024] The official implementation of our ICLR2024 paper "AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language M…☆345Updated 5 months ago
- UP-TO-DATE LLM Watermark paper. 🔥🔥🔥☆342Updated 6 months ago
- An easy-to-use Python framework to generate adversarial jailbreak prompts.☆673Updated 3 months ago
- MarkLLM: An Open-Source Toolkit for LLM Watermarking.(EMNLP 2024 Demo)☆418Updated 2 weeks ago
- A survey on harmful fine-tuning attack for large language model☆184Updated last week
- [ICML 2024] TrustLLM: Trustworthiness in Large Language Models☆574Updated 3 months ago
- Accepted by IJCAI-24 Survey Track☆207Updated 10 months ago
- LLM hallucination paper list☆318Updated last year
- Awesome RL-based LLM Reasoning☆526Updated last month
- ShieldLM: Empowering LLMs as Aligned, Customizable and Explainable Safety Detectors [EMNLP 2024 Findings]☆198Updated 9 months ago
- This is the repository for the Tool Learning survey.☆395Updated last month
- Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, exciting jailbreak methods on LLMs. It contains papers, codes, data…☆763Updated last week
- Official github repo for SafetyBench, a comprehensive benchmark to evaluate LLMs' safety. [ACL 2024]☆225Updated last year
- Paper list for Efficient Reasoning.☆509Updated this week
- JailbreakBench: An Open Robustness Benchmark for Jailbreaking Language Models [NeurIPS 2024 Datasets and Benchmarks Track]☆359Updated 2 months ago