sshh12 / llm_backdoorLinks
Experimental tools to backdoor large language models by re-writing their system prompts at a raw parameter level. This allows you to potentially execute offline remote code execution without running any actual code on the victim's machine or thwart LLM-based fraud/moderation systems.
☆174Updated 3 months ago
Alternatives and similar repositories for llm_backdoor
Users that are interested in llm_backdoor are comparing it to the libraries listed below
Sorting:
- ComPromptMized: Unleashing Zero-click Worms that Target GenAI-Powered Applications☆203Updated last year
- Use LLMs for document ranking☆144Updated 3 months ago
- Lightweight LLM Interaction Framework☆296Updated this week
- Codebase of https://arxiv.org/abs/2410.14923☆48Updated 8 months ago
- This repository contains various attack against Large Language Models.☆109Updated last year
- A productionized greedy coordinate gradient (GCG) attack tool for large language models (LLMs)☆122Updated 6 months ago
- Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks☆71Updated last month
- A MCP server for using Semgrep to scan code for security vulnerabilities.☆220Updated this week
- Repo with random useful scripts, utilities, prompts and stuff☆114Updated last week
- ☆43Updated this week
- Using Agents To Automate Pentesting☆281Updated 5 months ago
- A list of curated resources for people interested in AI Red Teaming, Jailbreaking, and Prompt Injection☆238Updated 2 months ago
- A Completely Modular LLM Reverse Engineering, Red Teaming, and Vulnerability Research Framework.☆46Updated 8 months ago
- Red-Teaming Language Models with DSPy☆202Updated 5 months ago
- A powerful tool for automated LLM fuzzing. It is designed to help developers and security researchers identify and mitigate potential jai…☆638Updated this week
- A very simple open source implementation of Google's Project Naptime☆160Updated 3 months ago
- A YAML based format for describing tools to LLMs, like man pages but for robots!☆75Updated 2 months ago
- A sandbox environment designed for loading, running and profiling a wide range of files, including machine learning models, ELFs, Pickle,…☆323Updated this week
- The LLM Red Teaming Framework☆512Updated last week
- Dropbox LLM Security research code and results☆228Updated last year
- A Model Context Protocol (MCP) server for querying the VirusTotal API.☆68Updated 4 months ago
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆92Updated 3 months ago
- A Model Context Protocol (MCP) server for querying the CVE-Search API☆34Updated 2 months ago
- https://arxiv.org/abs/2412.02776☆59Updated 7 months ago
- Top 10 for Agentic AI (AI Agent Security) serves as the core for OWASP and CSA Red teaming work☆119Updated last month
- A benchmark for prompt injection detection systems.☆122Updated 2 months ago
- A utility to inspect, validate, sign and verify machine learning model files.☆57Updated 5 months ago
- ☆54Updated last week
- OpenShield is a new generation security layer for AI models☆80Updated this week
- Every practical and proposed defense against prompt injection.☆495Updated 4 months ago