Trust4AI / ASTRALLinks
Automated Safety Testing of Large Language Models
☆16Updated 5 months ago
Alternatives and similar repositories for ASTRAL
Users that are interested in ASTRAL are comparing it to the libraries listed below
Sorting:
- ☆75Updated 8 months ago
- LLMs can be Dangerous Reasoners: Analyzing-based Jailbreak Attack on Large Language Models☆21Updated last month
- The jailbreak-evaluation is an easy-to-use Python package for language model jailbreak evaluation.☆24Updated 8 months ago
- ☆69Updated last year
- CyberGym is a large-scale, high-quality cybersecurity evaluation framework designed to rigorously assess the capabilities of AI agents on…☆45Updated 3 weeks ago
- An Execution Isolation Architecture for LLM-Based Agentic Systems☆83Updated 5 months ago
- CS-Eval is a comprehensive evaluation suite for fundamental cybersecurity models or large language models' cybersecurity ability.☆43Updated 7 months ago
- [NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.