ZiyueWang25 / llm-security-challenge
Can Large Language Models Solve Security Challenges? We test LLMs' ability to interact and break out of shell environments using the OverTheWire wargames environment, showing the models' surprising ability to do action-oriented cyberexploits in shell environments
☆12Updated last year
Alternatives and similar repositories for llm-security-challenge:
Users that are interested in llm-security-challenge are comparing it to the libraries listed below
- Whispers in the Machine: Confidentiality in LLM-integrated Systems☆35Updated last month
- 🤖🛡️🔍🔒🔑 Tiny package designed to support red teams and penetration testers in exploiting large language model AI solutions.☆23Updated 11 months ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆30Updated 10 months ago
- Agent Security Bench (ASB)☆76Updated 3 weeks ago
- PAL: Proxy-Guided Black-Box Attack on Large Language Models☆50Updated 8 months ago
- LLM security and privacy☆48Updated 6 months ago
- ☆54Updated 4 months ago
- LLM Self Defense: By Self Examination, LLMs know they are being tricked☆32Updated 11 months ago
- ☆59Updated 9 months ago
- Code to break Llama Guard☆31Updated last year
- This repository provides a benchmark for prompt Injection attacks and defenses☆188Updated last week
- ☆67Updated last month
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆109Updated last year
- An Execution Isolation Architecture for LLM-Based Agentic Systems☆70Updated 2 months ago
- Finding trojans in aligned LLMs. Official repository for the competition hosted at SaTML 2024.☆111Updated 10 months ago
- AutoDefense: Multi-Agent LLM Defense against Jailbreak Attacks☆44Updated 10 months ago
- ☆59Updated 5 months ago
- The D-CIPHER and NYU CTF baseline LLM Agents built for NYU CTF Bench☆65Updated 2 weeks ago
- ☆20Updated last year
- [ICML 2024] COLD-Attack: Jailbreaking LLMs with Stealthiness and Controllability☆150Updated 4 months ago
- [NeurIPS 2024] Official implementation for "AgentPoison: Red-teaming LLM Agents via Memory or Knowledge Base Backdoor Poisoning"☆114Updated 2 weeks ago
- LLM Platform Security: Applying a Systematic Evaluation Framework to OpenAI's ChatGPT Plugins☆25Updated 8 months ago
- A prompt injection game to collect data for robust ML research☆55Updated 2 months ago
- ☆93Updated last month
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆130Updated 3 weeks ago
- Papers about red teaming LLMs and Multimodal models.☆111Updated 5 months ago
- [ICLR 2025] Dissecting Adversarial Robustness of Multimodal LM Agents☆80Updated 2 months ago
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".☆102Updated last year
- Fine-tuning base models to build robust task-specific models☆29Updated last year
- ☆52Updated 2 months ago