aiverify-foundation / moonshot
Moonshot - A simple and modular tool to evaluate and red-team any LLM application.
☆235Updated this week
Alternatives and similar repositories for moonshot
Users that are interested in moonshot are comparing it to the libraries listed below
Sorting:
- AI Verify☆8Updated this week
- Contains all assets to run with Moonshot Library (Connectors, Datasets and Metrics)☆33Updated this week
- A Comprehensive Assessment of Trustworthiness in GPT Models☆290Updated 8 months ago
- This repository provides a benchmark for prompt Injection attacks and defenses☆197Updated 2 weeks ago
- ☆9Updated 3 months ago
- Collection of evals for Inspect AI☆131Updated this week
- ☆43Updated 9 months ago
- Red-Teaming Language Models with DSPy☆192Updated 3 months ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆369Updated last year
- A tool for evaluating LLMs☆419Updated last year
- The fastest Trust Layer for AI Agents☆133Updated 2 months ago
- Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs☆246Updated 11 months ago
- ⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs☆382Updated last year
- Fiddler Auditor is a tool to evaluate language models.☆179Updated last year
- AutoDefense: Multi-Agent LLM Defense against Jailbreak Attacks☆45Updated 11 months ago
- A benchmark for prompt injection detection systems.☆110Updated this week
- Papers about red teaming LLMs and Multimodal models.☆115Updated 5 months ago
- Attribute (or cite) statements generated by LLMs back to in-context information.☆231Updated 7 months ago
- Automatically evaluate your LLMs in Google Colab☆625Updated last year
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆90Updated this week
- Sample notebooks and prompts for LLM evaluation☆126Updated last week
- Inspect: A framework for large language model evaluations☆938Updated this week
- [NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.☆156Updated last month
- A Dynamic Environment to Evaluate Attacks and Defenses for LLM Agents.☆154Updated last week
- Dataset for the Tensor Trust project☆40Updated last year
- ☆100Updated 2 months ago
- This repository stems from our paper, “Cataloguing LLM Evaluations”, and serves as a living, collaborative catalogue of LLM evaluation fr…☆17Updated last year
- Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [ICLR 2025]☆304Updated 3 months ago
- Every practical and proposed defense against prompt injection.☆456Updated 2 months ago
- The Rule-based Retrieval package is a Python package that enables you to create and manage Retrieval Augmented Generation (RAG) applicati…☆238Updated 7 months ago