greshake / llm-security
New ways of breaking app-integrated LLMs
☆1,930Updated last year
Alternatives and similar repositories for llm-security
Users that are interested in llm-security are comparing it to the libraries listed below
Sorting:
- LLM Prompt Injection Detector☆1,273Updated 9 months ago
- Universal and Transferable Attacks on Aligned Language Models☆3,913Updated 9 months ago
- ☆549Updated 5 months ago
- A curated list of useful resources that cover Offensive AI.☆1,223Updated 2 weeks ago
- Dropbox LLM Security research code and results☆225Updated 11 months ago
- Minimal Python library to connect to LLMs (OpenAI, Anthropic, Google, Groq, Reka, Together, AI21, Cohere, Aleph Alpha, HuggingfaceHub), w…☆772Updated last month
- OWASP Foundation Web Respository☆734Updated this week
- A curation of awesome tools, documents and projects about LLM Security.☆1,217Updated last month
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆369Updated last year
- ☆1,472Updated 2 years ago
- A curated list of awesome security tools, experimental case or other interesting things with LLM or GPT.☆594Updated 4 months ago
- Persuasive Jailbreaker: we can persuade LLMs to jailbreak them!☆299Updated 7 months ago
- Every practical and proposed defense against prompt injection.☆456Updated 2 months ago
- A curated list of large language model tools for cybersecurity research.☆454Updated last year
- The Security Toolkit for LLM Interactions☆1,673Updated this week
- Adding guardrails to large language models.☆4,923Updated last week
- Simple UI for LLM Model Finetuning☆2,061Updated last year
- a prompt injection scanner for custom LLM applications☆785Updated 2 months ago
- An LLM-based autonomous agent controlling real-world applications via RESTful APIs☆1,366Updated 11 months ago
- NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.☆4,725Updated this week
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆380Updated last year
- Make your GenAI Apps Safe & Secure Test & harden your system prompt☆475Updated 7 months ago
- HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusal☆643Updated 9 months ago
- Official repo for GPTFUZZER : Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts☆488Updated 7 months ago
- Protection against Model Serialization Attacks☆478Updated this week
- CodeTF: One-stop Transformer Library for State-of-the-art Code LLM☆1,477Updated 2 weeks ago
- The Python Risk Identification Tool for generative AI (PyRIT) is an open source framework built to empower security professionals and eng…☆2,477Updated last week
- A collection of modular datasets generated by GPT-4, General-Instruct - Roleplay-Instruct - Code-Instruct - and Toolformer☆1,630Updated last year
- Agent techniques to augment your LLM and push it beyong its limits☆1,576Updated 11 months ago
- Prompt Engineering | Prompt Versioning | Use GPT or other prompt based models to get structured output. Join our discord for Prompt-Engin…☆3,695Updated 3 months ago