sunblaze-ucb / cybergymLinks
CyberGym is a large-scale, high-quality cybersecurity evaluation framework designed to rigorously assess the capabilities of AI agents on real-world vulnerability analysis tasks.
☆21Updated 2 weeks ago
Alternatives and similar repositories for cybergym
Users that are interested in cybergym are comparing it to the libraries listed below
Sorting:
- future-proof vulnerability detection benchmark, based on CVEs in open-source repos☆56Updated last week
- The D-CIPHER and NYU CTF baseline LLM Agents built for NYU CTF Bench☆81Updated 2 months ago
- ☆112Updated last week
- ☆55Updated last month
- CVE-Bench: A Benchmark for AI Agents’ Ability to Exploit Real-World Web Application Vulnerabilities☆58Updated last week
- An Execution Isolation Architecture for LLM-Based Agentic Systems☆82Updated 4 months ago
- ☆65Updated 5 months ago
- Repository for "SecurityEval Dataset: Mining Vulnerability Examples to Evaluate Machine Learning-Based Code Generation Techniques" publis…☆71Updated last year
- ☆43Updated 8 months ago
- ☆26Updated last year
- ☆34Updated 7 months ago
- LLM security and privacy☆48Updated 8 months ago
- SecLLMHolmes is a generalized, fully automated, and scalable framework to systematically evaluate the performance (i.e., accuracy and rea…☆58Updated last month
- 🪐 A Database of Existing Security Vulnerabilities Patches to Enable Evaluation of Techniques (single-commit; multi-language)☆40Updated 2 months ago
- An autonomous LLM-agent for large-scale, repository-level code auditing☆82Updated 2 weeks ago
- Challenge Problem #1 - Linux Kernel (NOTE: This code does not reflect the active state of what will be used at competition time, please r…☆53Updated last year
- VulZoo: A Comprehensive Vulnerability Intelligence Dataset (ASE 2024 Demo)☆51Updated 3 months ago
- [CCS'24] An LLM-based, fully automated fuzzing tool for option combination testing.☆82Updated 2 months ago
- 🤖🛡️🔍🔒🔑 Tiny package designed to support red teams and penetration testers in exploiting large language model AI solutions.☆23Updated last year
- CS-Eval is a comprehensive evaluation suite for fundamental cybersecurity models or large language models' cybersecurity ability.☆43Updated 6 months ago
- ☆65Updated 11 months ago
- General research for Dreadnode☆23Updated last year
- ☆72Updated 7 months ago
- A collection of prompt injection mitigation techniques.☆23Updated last year
- ☆116Updated 11 months ago
- ☆41Updated 8 months ago
- CyberBench: A Multi-Task Cyber LLM Benchmark☆17Updated last month
- Code used to run the platform for the LLM CTF colocated with SaTML 2024☆26Updated last year
- PAL: Proxy-Guided Black-Box Attack on Large Language Models☆51Updated 10 months ago
- using ML models for red teaming☆43Updated last year