corca-ai / LLMFuzzAgentLinks
[Corca / ML] Automatically solved Gandalf AI with LLM
☆50Updated last year
Alternatives and similar repositories for LLMFuzzAgent
Users that are interested in LLMFuzzAgent are comparing it to the libraries listed below
Sorting:
- Red-Teaming Language Models with DSPy☆198Updated 4 months ago
- Curation of prompts that are known to be adversarial to large language models☆179Updated 2 years ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆112Updated last year
- Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platfor…☆29Updated last year
- Dropbox LLM Security research code and results☆227Updated last year
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆91Updated 2 months ago
- Payloads for Attacking Large Language Models☆90Updated 3 weeks ago
- 🧠 LLMFuzzer - Fuzzing Framework for Large Language Models 🧠 LLMFuzzer is the first open-source fuzzing framework specifically designed …☆282Updated last year
- HoneyAgents is a PoC demo of an AI-driven system that combines honeypots with autonomous AI agents to detect and mitigate cyber threats. …☆52Updated last year
- A guide to LLM hacking: fundamentals, prompt injection, offense, and defense☆159Updated 2 years ago
- ☆22Updated last year
- A benchmark for prompt injection detection systems.☆120Updated last month
- Fiddler Auditor is a tool to evaluate language models.☆183Updated last year
- My inputs for the LLM Gandalf made by Lakera☆43Updated last year
- This project investigates the security of large language models by performing binary classification of a set of input prompts to discover…☆40Updated last year
- Track the progress of LLM context utilisation☆54Updated 2 months ago
- ☆65Updated 5 months ago
- Approximation of the Claude 3 tokenizer by inspecting generation stream☆131Updated 11 months ago
- Masked Python SDK wrapper for OpenAI API. Use public LLM APIs securely.☆116Updated 2 years ago
- A text embedding viewer for the Jupyter environment☆20Updated last year
- Risks and targets for assessing LLMs & LLM vulnerabilities☆30Updated last year
- Uses the ChatGPT model to determine if a user-supplied question is safe and filter out dangerous questions☆48Updated 2 years ago
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆203Updated last year
- Evaluate your LLM apps, RAG pipeline, any generated text, and more!☆1Updated last year
- A library for red-teaming LLM applications with LLMs.☆26Updated 8 months ago
- ☆116Updated 2 weeks ago
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆184Updated 2 months ago
- Security measure for agentic LLMs using a council of AIs moderted by a veto system. The council judges an agent's actions outputs based o…☆38Updated 2 years ago
- Security Threats related with MCP (Model Context Protocol), MCP Servers and more☆27Updated 2 months ago
- Project LLM Verification Standard☆44Updated last month