corca-ai / LLMFuzzAgent
[Corca / ML] Automatically solved Gandalf AI with LLM
☆50Updated last year
Alternatives and similar repositories for LLMFuzzAgent
Users that are interested in LLMFuzzAgent are comparing it to the libraries listed below
Sorting:
- Dropbox LLM Security research code and results☆225Updated 11 months ago
- Red-Teaming Language Models with DSPy☆192Updated 3 months ago
- Payloads for Attacking Large Language Models☆85Updated 10 months ago
- Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platfor…☆29Updated last year
- ☆100Updated 2 months ago
- Curation of prompts that are known to be adversarial to large language models☆179Updated 2 years ago
- source for llmsec.net☆15Updated 9 months ago
- Fiddler Auditor is a tool to evaluate language models.☆179Updated last year
- 🧠 LLMFuzzer - Fuzzing Framework for Large Language Models 🧠 LLMFuzzer is the first open-source fuzzing framework specifically designed …☆278Updated last year
- Uses the ChatGPT model to determine if a user-supplied question is safe and filter out dangerous questions☆46Updated 2 years ago
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆91Updated last month
- [Corca / OR] Solver for Multi-dimensional Multi-demand Quadratic Knapsack Problems☆12Updated 3 years ago
- Lakera - ChatGPT Data Leak Protection☆22Updated 10 months ago
- Evaluate your LLM apps, RAG pipeline, any generated text, and more!☆1Updated last year
- Guard your LangChain applications against prompt injection with Lakera ChainGuard.☆22Updated 2 months ago
- A guide to LLM hacking: fundamentals, prompt injection, offense, and defense☆152Updated 2 years ago
- Code to break Llama Guard☆31Updated last year
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆202Updated last year
- A toolkit for detecting and protecting against vulnerabilities in Large Language Models (LLMs).☆136Updated last year
- A library for red-teaming LLM applications with LLMs.☆26Updated 7 months ago
- A Python-based tool that monitors dark web sources for mentions of specific organizations for Threat Monitoring.☆17Updated last month
- Security Threats related with MCP (Model Context Protocol), MCP Servers and more☆24Updated 3 weeks ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆369Updated last year
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".☆102Updated last year
- Project LLM Verification Standard☆43Updated last year
- Stanford CRFM's initiative to assess potential compliance with the draft EU AI Act☆94Updated last year
- Masked Python SDK wrapper for OpenAI API. Use public LLM APIs securely.☆117Updated 2 years ago
- A collection of prompt injection mitigation techniques.☆22Updated last year
- Sphynx Hallucination Induction☆54Updated 3 months ago
- Framework for LLM evaluation, guardrails and security☆112Updated 8 months ago