Azure / PyRITLinks
The Python Risk Identification Tool for generative AI (PyRIT) is an open source framework built to empower security professionals and engineers to proactively identify risks in generative AI systems.
☆3,380Updated this week
Alternatives and similar repositories for PyRIT
Users that are interested in PyRIT are comparing it to the libraries listed below
Sorting:
- the LLM vulnerability scanner☆6,892Updated this week
- OWASP Top 10 for Large Language Model Apps (Part of the GenAI Security Project)☆1,066Updated this week
- Set of tools to assess and improve LLM security.☆4,011Updated this week
- The Security Toolkit for LLM Interactions☆2,476Updated last month
- LLM Prompt Injection Detector☆1,415Updated last year
- Protection against Model Serialization Attacks☆642Updated 2 months ago
- Make your GenAI Apps Safe & Secure Test & harden your system prompt☆618Updated last week
- Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪☆1,761Updated last week
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆452Updated 2 years ago
- A curated list of large language model tools for cybersecurity research.☆481Updated last year
- AI Red Teaming playground labs to run AI Red Teaming trainings including infrastructure.☆1,808Updated 2 weeks ago
- Every practical and proposed defense against prompt injection.☆624Updated 11 months ago
- a security scanner for custom LLM applications☆1,119Updated 2 months ago
- A curation of awesome tools, documents and projects about LLM Security.☆1,524Updated 5 months ago
- a CLI that provides a generic automation layer for assessing the security of ML models☆907Updated 6 months ago
- An offensive/defense security toolset for discovery, recon and ethical assessment of AI Agents☆1,111Updated last month
- Microsoft Security Copilot is a generative AI-powered security solution that helps increase the efficiency and capabilities of defenders …☆602Updated 5 months ago
- New ways of breaking app-integrated LLMs☆2,049Updated 6 months ago
- AttackGen is a cybersecurity incident response testing tool that leverages the power of large language models and the comprehensive MITRE…☆1,207Updated last month
- Test Software for the Characterization of AI Technologies☆277Updated this week
- 🧠 LLMFuzzer - Fuzzing Framework for Large Language Models 🧠 LLMFuzzer is the first open-source fuzzing framework specifically designed …☆338Updated last year
- DeepTeam is a framework to red team LLMs and LLM systems.☆1,263Updated last week
- A powerful tool for automated LLM fuzzing. It is designed to help developers and security researchers identify and mitigate potential jai…☆1,185Updated 2 months ago
- LLM powered fuzzing via OSS-Fuzz.☆1,353Updated last week
- Zero shot vulnerability discovery using LLMs☆2,476Updated last year
- Modern CLI for exploring vulnerability data with powerful search, filtering, and analysis capabilities.☆2,353Updated 2 weeks ago
- A unified evaluation framework for large language models☆2,773Updated 2 weeks ago
- Prompt Injection Primer for Engineers☆546Updated 2 years ago
- A collection of awesome resources related AI security☆533Updated this week
- HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusal☆841Updated last year