haizelabs / BEAST-implementation
☆16Updated 10 months ago
Alternatives and similar repositories for BEAST-implementation:
Users that are interested in BEAST-implementation are comparing it to the libraries listed below
- ☆64Updated 2 months ago
- A utility to inspect, validate, sign and verify machine learning model files.☆55Updated last month
- Tree of Attacks (TAP) Jailbreaking Implementation☆105Updated last year
- General research for Dreadnode☆20Updated 9 months ago
- A YAML based format for describing tools to LLMs, like man pages but for robots!☆67Updated last month
- Red-Teaming Language Models with DSPy☆175Updated last month
- ☆87Updated last month
- A subset of jailbreaks automatically discovered by the Haize Labs haizing suite.☆89Updated 9 months ago
- future-proof vulnerability detection benchmark, based on CVEs in open-source repos☆51Updated this week
- Sphynx Hallucination Induction☆53Updated 2 months ago
- A collection of prompt injection mitigation techniques.☆20Updated last year
- Data Scientists Go To Jupyter☆62Updated 3 weeks ago
- Thorn in a HaizeStack test for evaluating long-context adversarial robustness.☆26Updated 7 months ago
- A library for red-teaming LLM applications with LLMs.☆25Updated 5 months ago
- CLI and API server for https://github.com/dreadnode/robopages☆32Updated last week
- Small tools to assist with using Large Language Models☆11Updated last year
- Codebase of https://arxiv.org/abs/2410.14923☆44Updated 5 months ago
- https://arxiv.org/abs/2412.02776☆51Updated 3 months ago
- ☆19Updated last year
- [IJCAI 2024] Imperio is an LLM-powered backdoor attack. It allows the adversary to issue language-guided instructions to control the vict…☆41Updated last month
- Code to break Llama Guard☆31Updated last year
- ☆31Updated 4 months ago
- using ML models for red teaming☆43Updated last year
- A framework-less approach to robust agent development.☆156Updated this week
- ☆30Updated 6 months ago
- A Completely Modular LLM Reverse Engineering, Red Teaming, and Vulnerability Research Framework.☆46Updated 4 months ago
- CompChomper is a framework for measuring how LLMs perform at code completion.☆16Updated last month
- source code for the offsecml framework☆38Updated 9 months ago
- CodeSage: Code Representation Learning At Scale (ICLR 2024)☆99Updated 5 months ago
- A prompt injection game to collect data for robust ML research☆54Updated 2 months ago