Giskard-AI / awesome-ai-safetyLinks
π A curated list of papers & technical articles on AI Quality & Safety
β199Updated 9 months ago
Alternatives and similar repositories for awesome-ai-safety
Users that are interested in awesome-ai-safety are comparing it to the libraries listed below
Sorting:
- Fiddler Auditor is a tool to evaluate language models.β188Updated last year
- β271Updated 11 months ago
- Stanford CRFM's initiative to assess potential compliance with the draft EU AI Actβ93Updated 2 years ago
- Red-Teaming Language Models with DSPyβ250Updated 11 months ago
- π Datasets and models for instruction-tuningβ238Updated 2 years ago
- RAGElo is a set of tools that helps you selecting the best RAG-based LLM agents by using an Elo rankerβ126Updated 2 months ago
- β261Updated 9 months ago
- A tool for evaluating LLMsβ427Updated last year
- Mixing Language Models with Self-Verification and Meta-Verificationβ111Updated last year
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to aβ¦β449Updated last year
- β339Updated 2 years ago
- π A curated list of resources dedicated to synthetic dataβ140Updated 3 years ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absoluteβ¦β51Updated last year
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.β116Updated 5 months ago
- A joint community effort to create one central leaderboard for LLMs.β308Updated last year
- An open-source compliance-centered evaluation framework for Generative AI modelsβ178Updated 3 weeks ago
- π¦ Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data β¦β212Updated last week
- β42Updated last year
- Open Implementations of LLM Analysesβ107Updated last year
- Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuningβ46Updated 2 years ago
- A comprehensive guide to LLM evaluation methods designed to assist in identifying the most suitable evaluation techniques for various useβ¦β172Updated last week
- AI Verifyβ40Updated last week
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.β118Updated last week
- The Foundation Model Transparency Indexβ85Updated last month
- Toolkit for attaching, training, saving and loading of new heads for transformer modelsβ294Updated 10 months ago
- Evaluating LLMs with fewer examplesβ169Updated last year
- A curated list of awesome synthetic data tools (open source and commercial).β231Updated 2 years ago
- Curation of prompts that are known to be adversarial to large language modelsβ188Updated 2 years ago
- Backtracing: Retrieving the Cause of the Query, EACL 2024 Long Paper, Findings.β92Updated last year
- A curated list of awesome publications and researchers on prompting framework updated and maintained by The Intelligent System Security (β¦β85Updated last year