ZiyueWang25 / llm-security-challengeLinks
Can Large Language Models Solve Security Challenges? We test LLMs' ability to interact and break out of shell environments using the OverTheWire wargames environment, showing the models' surprising ability to do action-oriented cyberexploits in shell environments
☆13Updated 2 years ago
Alternatives and similar repositories for llm-security-challenge
Users that are interested in llm-security-challenge are comparing it to the libraries listed below
Sorting:
- Whispers in the Machine: Confidentiality in Agentic Systems☆41Updated last week
- LLM security and privacy☆52Updated last year
- [NeurIPS 2024] Official implementation for "AgentPoison: Red-teaming LLM Agents via Memory or Knowledge Base Backdoor Poisoning"☆181Updated 8 months ago
- Risks and targets for assessing LLMs & LLM vulnerabilities☆33Updated last year
- AutoDefense: Multi-Agent LLM Defense against Jailbreak Attacks☆61Updated 6 months ago
- CyberGym is a large-scale, high-quality cybersecurity evaluation framework designed to rigorously assess the capabilities of AI agents on…☆100Updated 2 months ago
- Code used to run the platform for the LLM CTF colocated with SaTML 2024☆28Updated last year
- 🤖🛡️🔍🔒🔑 Tiny package designed to support red teams and penetration testers in exploiting large language model AI solutions.☆27Updated last year
- LLM Platform Security: Applying a Systematic Evaluation Framework to OpenAI's ChatGPT Plugins☆29Updated last year
- [NeurIPS'24] RedCode: Risky Code Execution and Generation Benchmark for Code Agents☆60Updated last month
- An Execution Isolation Architecture for LLM-Based Agentic Systems☆100Updated 10 months ago
- ☆108Updated 2 weeks ago
- ☆99Updated last year
- This repository provides a benchmark for prompt injection attacks and defenses in LLMs☆361Updated last month
- Repository for "SecurityEval Dataset: Mining Vulnerability Examples to Evaluate Machine Learning-Based Code Generation Techniques" publis…☆83Updated 2 years ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆443Updated last year
- Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [ICLR 2025]☆368Updated 10 months ago
- Implementation of BEAST adversarial attack for language models (ICML 2024)☆92Updated last year
- Code to break Llama Guard☆32Updated 2 years ago
- Agent Security Bench (ASB)☆155Updated last month
- ☆124Updated last year
- PAL: Proxy-Guided Black-Box Attack on Large Language Models☆55Updated last year
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆379Updated 2 weeks ago
- Universal Robustness Evaluation Toolkit (for Evasion)☆32Updated 3 months ago
- ☆54Updated last year
- Package to optimize Adversarial Attacks against (Large) Language Models with Varied Objectives☆70Updated last year
- The D-CIPHER and NYU CTF baseline LLM Agents built for NYU CTF Bench☆110Updated last month
- A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.☆91Updated last year
- Papers about red teaming LLMs and Multimodal models.☆157Updated 6 months ago
- Persuasive Jailbreaker: we can persuade LLMs to jailbreak them!☆341Updated 2 months ago