sshh12 / llm_backdoorLinks
Experimental tools to backdoor large language models by re-writing their system prompts at a raw parameter level. This allows you to potentially execute offline remote code execution without running any actual code on the victim's machine or thwart LLM-based fraud/moderation systems.
☆186Updated this week
Alternatives and similar repositories for llm_backdoor
Users that are interested in llm_backdoor are comparing it to the libraries listed below
Sorting:
- ☆44Updated last week
- Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks☆86Updated 4 months ago
- Here Comes the AI Worm: Preventing the Propagation of Adversarial Self-Replicating Prompts Within GenAI Ecosystems☆211Updated last month
- A Completely Modular LLM Reverse Engineering, Red Teaming, and Vulnerability Research Framework.☆51Updated 11 months ago
- Code for the paper "Defeating Prompt Injections by Design"☆125Updated 3 months ago
- Lightweight LLM Interaction Framework☆381Updated this week
- Codebase of https://arxiv.org/abs/2410.14923☆51Updated 11 months ago
- A productionized greedy coordinate gradient (GCG) attack tool for large language models (LLMs)☆140Updated 9 months ago
- Repo with random useful scripts, utilities, prompts and stuff☆171Updated 2 months ago
- Code snippets to reproduce MCP tool poisoning attacks.☆181Updated 6 months ago
- Use LLMs for document ranking☆148Updated 5 months ago
- This repository contains various attack against Large Language Models.☆114Updated last year
- We present MAPTA, a multi-agent system for autonomous web application security assessment that combines large language model orchestratio…☆65Updated last month
- LLM | Security | Operations in one github repo with good links and pictures.☆58Updated 9 months ago
- A YAML based format for describing tools to LLMs, like man pages but for robots!☆77Updated 5 months ago
- A powerful tool for automated LLM fuzzing. It is designed to help developers and security researchers identify and mitigate potential jai…☆775Updated 2 months ago
- A utility to inspect, validate, sign and verify machine learning model files.☆58Updated 8 months ago
- https://arxiv.org/abs/2412.02776☆62Updated 10 months ago
- Using Agents To Automate Pentesting☆301Updated 8 months ago
- Dropbox LLM Security research code and results☆235Updated last year
- A list of curated resources for people interested in AI Red Teaming, Jailbreaking, and Prompt Injection☆330Updated 5 months ago
- All-in-one offensive security toolbox with AI agent and MCP architecture. Integrates tools like Nmap, Metasploit, FFUF, SQLMap. Enables p…☆425Updated 4 months ago
- A knowledge source about TTPs used to target GenAI-based systems, copilots and agents☆122Updated last week
- A sandbox environment designed for loading, running and profiling a wide range of files, including machine learning models, ELFs, Pickle,…☆330Updated last week
- Cybersecurity Intelligent Pentesting Helper for Ethical Researcher (CIPHER). Fine tuned LLM for penetration testing guidance based on wri…☆31Updated 9 months ago
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆96Updated 5 months ago
- ☆76Updated this week
- A very simple open source implementation of Google's Project Naptime☆169Updated 6 months ago
- ☆68Updated 2 months ago
- Red-Teaming Language Models with DSPy☆216Updated 7 months ago