lve-org / lve
A repository of Language Model Vulnerabilities and Exposures (LVEs).
☆109Updated last year
Alternatives and similar repositories for lve:
Users that are interested in lve are comparing it to the libraries listed below
- ☆98Updated 2 months ago
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆142Updated this week
- TaskTracker is an approach to detecting task drift in Large Language Models (LLMs) by analysing their internal activations. It provides a…☆54Updated last month
- Risks and targets for assessing LLMs & LLM vulnerabilities☆30Updated 11 months ago
- future-proof vulnerability detection benchmark, based on CVEs in open-source repos☆52Updated last week
- A benchmark for prompt injection detection systems.☆106Updated last week
- Red-Teaming Language Models with DSPy☆188Updated 2 months ago
- Dropbox LLM Security research code and results☆224Updated 11 months ago
- [NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.☆152Updated last month
- A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.☆66Updated last year
- Tree of Attacks (TAP) Jailbreaking Implementation☆107Updated last year
- WMDP is a LLM proxy benchmark for hazardous knowledge in bio, cyber, and chemical security. We also release code for RMU, an unlearning m…☆113Updated last year
- This repository provides a benchmark for prompt Injection attacks and defenses☆192Updated this week
- 🧠 LLMFuzzer - Fuzzing Framework for Large Language Models 🧠 LLMFuzzer is the first open-source fuzzing framework specifically designed …☆274Updated last year
- The official implementation of our pre-print paper "Automatic and Universal Prompt Injection Attacks against Large Language Models".☆46Updated 6 months ago
- Code to break Llama Guard☆31Updated last year
- Papers about red teaming LLMs and Multimodal models.☆113Updated 5 months ago
- ☆128Updated last week
- Finding trojans in aligned LLMs. Official repository for the competition hosted at SaTML 2024.☆111Updated 10 months ago
- LLM security and privacy☆49Updated 6 months ago
- Implementation of BEAST adversarial attack for language models (ICML 2024)☆85Updated 11 months ago
- PAL: Proxy-Guided Black-Box Attack on Large Language Models☆50Updated 8 months ago
- ☆43Updated 9 months ago
- ATLAS tactics, techniques, and case studies data☆67Updated 2 weeks ago
- Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [ICLR 2025]☆299Updated 3 months ago
- The automated prompt injection framework for LLM-integrated applications.☆200Updated 7 months ago
- ☆32Updated 5 months ago
- TAP: An automated jailbreaking method for black-box LLMs☆165Updated 4 months ago
- A prompt injection game to collect data for robust ML research☆55Updated 3 months ago
- Guardrails for secure and robust agent development☆243Updated 2 weeks ago