ZiyueWang25 / llm-security-challengeLinks
Can Large Language Models Solve Security Challenges? We test LLMs' ability to interact and break out of shell environments using the OverTheWire wargames environment, showing the models' surprising ability to do action-oriented cyberexploits in shell environments
☆13Updated last year
Alternatives and similar repositories for llm-security-challenge
Users that are interested in llm-security-challenge are comparing it to the libraries listed below
Sorting:
- Whispers in the Machine: Confidentiality in Agentic Systems☆39Updated last month
- LLM security and privacy☆48Updated 9 months ago
- CyberGym is a large-scale, high-quality cybersecurity evaluation framework designed to rigorously assess the capabilities of AI agents on…☆44Updated 3 weeks ago
- 🤖🛡️🔍🔒🔑 Tiny package designed to support red teams and penetration testers in exploiting large language model AI solutions.☆23Updated last year
- Agent Security Bench (ASB)☆94Updated last month
- LLM Platform Security: Applying a Systematic Evaluation Framework to OpenAI's ChatGPT Plugins☆25Updated 11 months ago
- This repository provides a benchmark for prompt Injection attacks and defenses☆245Updated last month
- PAL: Proxy-Guided Black-Box Attack on Large Language Models☆51Updated 11 months ago
- An Execution Isolation Architecture for LLM-Based Agentic Systems☆83Updated 5 months ago
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆209Updated this week
- ☆75Updated 7 months ago
- Automated Safety Testing of Large Language Models☆16Updated 5 months ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆31Updated last year
- [NeurIPS 2024] Official implementation for "AgentPoison: Red-teaming LLM Agents via Memory or Knowledge Base Backdoor Poisoning"☆130Updated 3 months ago
- Official repo for GPTFUZZER : Red Teaming Large Language Models with Auto-Generated Jailbreak Prompts☆504Updated 9 months ago
- Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [ICLR 2025]☆320Updated 5 months ago
- [NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.☆162Updated 3 months ago
- AutoDefense: Multi-Agent LLM Defense against Jailbreak Attacks☆51Updated last month
- ☆67Updated last year
- Code used to run the platform for the LLM CTF colocated with SaTML 2024☆26Updated last year
- Code to break Llama Guard☆31Updated last year
- 🧠 LLMFuzzer - Fuzzing Framework for Large Language Models 🧠 LLMFuzzer is the first open-source fuzzing framework specifically designed …☆288Updated last year
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆393Updated last year
- Dataset for the Tensor Trust project