chziakas / redevalLinks
A library for red-teaming LLM applications with LLMs.
☆28Updated last year
Alternatives and similar repositories for redeval
Users that are interested in redeval are comparing it to the libraries listed below
Sorting:
- Red-Teaming Language Models with DSPy☆248Updated 10 months ago
- A prompt injection game to collect data for robust ML research☆65Updated 10 months ago
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆100Updated 8 months ago
- ☆66Updated 3 months ago
- The fastest Trust Layer for AI Agents☆144Updated 6 months ago
- ☆180Updated last week
- LLM security and privacy☆52Updated last year
- Official repository for the paper "ALERT: A Comprehensive Benchmark for Assessing Large Language Models’ Safety through Red Teaming"☆52Updated last year
- ☆29Updated 7 months ago
- Sphynx Hallucination Induction☆53Updated 10 months ago
- Code for the paper "Fishing for Magikarp"☆177Updated 7 months ago
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆112Updated last year
- The jailbreak-evaluation is an easy-to-use Python package for language model jailbreak evaluation.☆26Updated last year
- ☆108Updated 3 weeks ago
- Papers about red teaming LLMs and Multimodal models.☆158Updated 6 months ago
- ☆38Updated 6 months ago
- ☆34Updated last year
- AgentFence is an open-source platform for automatically testing AI agent security. It identifies vulnerabilities such as prompt injection…☆45Updated 9 months ago
- This project investigates the security of large language models by performing binary classification of a set of input prompts to discover…☆55Updated 2 years ago
- autoredteam: code for training models that automatically red team other language models☆15Updated 2 years ago
- Curation of prompts that are known to be adversarial to large language models☆186Updated 2 years ago
- ☆16Updated last year
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆195Updated 8 months ago
- Track the progress of LLM context utilisation☆55Updated 8 months ago
- ☆26Updated last year
- The official repository of the paper "On the Exploitability of Instruction Tuning".☆66Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆111Updated last year
- ☆86Updated last year
- 🤖🛡️🔍🔒🔑 Tiny package designed to support red teams and penetration testers in exploiting large language model AI solutions.☆27Updated last year
- ☆63Updated last week