aiverify-foundation / moonshot-dataLinks
Contains all assets to run with Moonshot Library (Connectors, Datasets and Metrics)
☆36Updated 2 weeks ago
Alternatives and similar repositories for moonshot-data
Users that are interested in moonshot-data are comparing it to the libraries listed below
Sorting:
- ☆42Updated last year
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆101Updated last week
- ☆34Updated 9 months ago
- Code for the paper "Fishing for Magikarp"☆163Updated 3 months ago
- ☆45Updated 4 months ago
- An open-source compliance-centered evaluation framework for Generative AI models☆161Updated last week
- ☆63Updated 3 weeks ago
- Red-Teaming Language Models with DSPy☆212Updated 6 months ago
- codebase release for EMNLP2023 paper publication☆19Updated 3 months ago
- Open Implementations of LLM Analyses☆106Updated 10 months ago
- ☆78Updated last week
- ☆39Updated 2 years ago
- Official repository for the paper "ALERT: A Comprehensive Benchmark for Assessing Large Language Models’ Safety through Red Teaming"☆45Updated 11 months ago
- [ICLR 2025] 🚀 CodeMMLU Evaluator: A framework for evaluating LM models on CodeMMLU MCQs benchmark.☆23Updated 4 months ago
- Ferret: Faster and Effective Automated Red Teaming with Reward-Based Scoring Technique☆18Updated last year
- Do-Not-Answer: A Dataset for Evaluating Safeguards in LLMs☆281Updated last year
- The Granite Guardian models are designed to detect risks in prompts and responses.☆102Updated 3 weeks ago
- ☆30Updated 2 months ago
- FBI: Finding Blindspots in LLM Evaluations with Interpretable Checklists☆29Updated 2 weeks ago
- Improving Alignment and Robustness with Circuit Breakers☆228Updated 11 months ago
- A simple evaluation of generative language models and safety classifiers.☆60Updated last year
- Code accompanying "How I learned to start worrying about prompt formatting".☆109Updated 2 months ago
- NeurIPS'24 - LLM Safety Landscape☆28Updated 6 months ago
- Evaluating LLMs with fewer examples☆160Updated last year
- [NeurIPS 2024] Goldfish Loss: Mitigating Memorization in Generative LLMs☆90Updated 9 months ago
- ☆38Updated 10 months ago
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆206Updated this week
- AuditNLG: Auditing Generative AI Language Modeling for Trustworthiness☆102Updated 7 months ago
- ☆74Updated last year
- Moonshot - A simple and modular tool to evaluate and red-team any LLM application.☆266Updated 3 weeks ago