aiverify-foundation / moonshot-dataLinks
Contains all assets to run with Moonshot Library (Connectors, Datasets and Metrics)
☆39Updated this week
Alternatives and similar repositories for moonshot-data
Users that are interested in moonshot-data are comparing it to the libraries listed below
Sorting:
- ☆35Updated last year
- Code for the paper "Fishing for Magikarp"☆176Updated 7 months ago
- The Granite Guardian models are designed to detect risks in prompts and responses.☆123Updated 2 months ago
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆114Updated this week
- ☆43Updated last year
- Official repository for the paper "ALERT: A Comprehensive Benchmark for Assessing Large Language Models’ Safety through Red Teaming"☆51Updated last year
- Moonshot - A simple and modular tool to evaluate and red-team any LLM application.☆294Updated last week
- Red-Teaming Language Models with DSPy☆247Updated 10 months ago
- ☆42Updated last year
- A simple evaluation of generative language models and safety classifiers.☆79Updated last week
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆212Updated this week
- FBI: Finding Blindspots in LLM Evaluations with Interpretable Checklists☆30Updated 4 months ago
- AI Verify☆39Updated this week
- Collection of evals for Inspect AI☆313Updated this week
- [ICLR 2025] 🚀 CodeMMLU Evaluator: A framework for evaluating LM models on CodeMMLU MCQs benchmark.☆28Updated 8 months ago
- Ferret: Faster and Effective Automated Red Teaming with Reward-Based Scoring Technique☆18Updated last year
- Reward Model framework for LLM RLHF☆61Updated 2 years ago
- ☆38Updated 2 years ago
- A Python library for guardrail models evaluation.☆29Updated 2 months ago
- A package dedicated for running benchmark agreement testing☆18Updated 3 months ago
- Evaluating LLMs with fewer examples☆170Updated last year
- Papers about red teaming LLMs and Multimodal models.☆157Updated 6 months ago
- ☆258Updated 8 months ago
- ☆49Updated last year
- AuditNLG: Auditing Generative AI Language Modeling for Trustworthiness☆101Updated 10 months ago
- Benchmarking Large Language Models☆104Updated 6 months ago
- Official Repo for CRMArena and CRMArena-Pro☆126Updated last month
- Codebase accompanying the Summary of a Haystack paper.☆79Updated last year
- Open Source Replication of Anthropic's Alignment Faking Paper☆52Updated 8 months ago
- An open-source compliance-centered evaluation framework for Generative AI models☆176Updated this week