corca-ai / awesome-llm-security
A curation of awesome tools, documents and projects about LLM Security.
β873Updated 3 weeks ago
Related projects: β
- Papers and resources related to the security and privacy of LLMs π€β393Updated last week
- A reading list for large models safety, security, and privacy (including Awesome LLM Security, Safety, etc.).β750Updated this week
- β357Updated last month
- A curated list of safety-related papers, articles, and resources focused on Large Language Models (LLMs). This repository aims to provideβ¦β844Updated this week
- LLM vulnerability scannerβ1,273Updated this week
- [ICML 2024] TrustLLM: Trustworthiness in Large Language Modelsβ432Updated 2 weeks ago
- Official repo for GPTFUZZER : Red Teaming Large Language Models with Auto-Generated Jailbreak Promptsβ366Updated 5 months ago
- This repository provides implementation to formalize and benchmark Prompt Injection attacks and defensesβ125Updated 2 weeks ago
- HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusalβ275Updated last month
- An easy-to-use Python framework to generate adversarial jailbreak prompts.β403Updated 2 weeks ago
- The papers are organized according to our survey: Evaluating Large Language Models: A Comprehensive Survey.β681Updated 4 months ago
- The official implementation of our ICLR2024 paper "AutoDAN: Generating Stealthy Jailbreak Prompts on Aligned Large Language Models".β203Updated last month
- Aligning Large Language Models with Human: A Surveyβ671Updated last year
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to aβ¦β293Updated 6 months ago
- An Open Robustness Benchmark for Jailbreaking Language Models [arXiv 2024]β169Updated last month
- TAP: An automated jailbreaking method for black-box LLMsβ106Updated 6 months ago
- Awesome-LLM-Robustness: a curated list of Uncertainty, Reliability and Robustness in Large Language Modelsβ627Updated 3 months ago
- OWASP Foundation Web Respositoryβ504Updated last week
- Must-read Papers on Knowledge Editing for Large Language Models.β829Updated 2 weeks ago
- We jailbreak GPT-3.5 Turboβs safety guardrails by fine-tuning it on only 10 adversarially designed examples, at a cost of less than $0.20β¦β219Updated 6 months ago
- Papers about red teaming LLMs and Multimodal models.β66Updated this week
- Every practical and proposed defense against prompt injection.β310Updated 3 months ago
- π§ LLMFuzzer - Fuzzing Framework for Large Language Models π§ LLMFuzzer is the first open-source fuzzing framework specifically designed β¦β218Updated 7 months ago
- An overview of LLMs for cybersecurity.β262Updated 3 months ago
- The official GitHub page for the survey paper "A Survey on Evaluation of Large Language Models".β1,382Updated 3 months ago
- LLM hallucination paper listβ268Updated 6 months ago
- List of papers on hallucination detection in LLMs.β561Updated last week
- A Comprehensive Assessment of Trustworthiness in GPT Modelsβ250Updated this week
- Reading list of hallucination in LLMs. Check out our new survey paper: "Sirenβs Song in the AI Ocean: A Survey on Hallucination in Large β¦β901Updated 2 weeks ago
- This is a collection of research papers for Self-Correcting Large Language Models with Automated Feedback.β386Updated 7 months ago