Giskard-AI / awesome-ai-safety
π A curated list of papers & technical articles on AI Quality & Safety
β169Updated last year
Alternatives and similar repositories for awesome-ai-safety:
Users that are interested in awesome-ai-safety are comparing it to the libraries listed below
- β262Updated 3 weeks ago
- Fiddler Auditor is a tool to evaluate language models.β175Updated 11 months ago
- Red-Teaming Language Models with DSPyβ168Updated last week
- Creating the tools and data sets necessary to evaluate vulnerabilities in LLMs.β23Updated last week
- π Datasets and models for instruction-tuningβ234Updated last year
- Stanford CRFM's initiative to assess potential compliance with the draft EU AI Actβ93Updated last year
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to aβ¦β333Updated 11 months ago
- β39Updated 6 months ago
- The Foundation Model Transparency Indexβ75Updated 8 months ago
- A framework-less approach to robust agent development.β154Updated this week
- SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Modelsβ493Updated 7 months ago
- RAGElo is a set of tools that helps you selecting the best RAG-based LLM agents by using an Elo rankerβ106Updated last week
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.β79Updated this week
- Mixing Language Models with Self-Verification and Meta-Verificationβ101Updated 2 months ago
- Automatically evaluate your LLMs in Google Colabβ592Updated 9 months ago
- β122Updated last week
- Web UI & Backend for Data Annotations in Ayaβ26Updated 11 months ago
- Benchmark various LLM Structured Output frameworks: Instructor, Mirascope, Langchain, LlamaIndex, Fructose, Marvin, Outlines, etc on taskβ¦β147Updated 4 months ago
- This is the reproduction repository for my π€ Hugging Face blog post on synthetic dataβ63Updated last year
- A curated list of awesome synthetic data tools (open source and commercial).β153Updated last year
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.β106Updated 5 months ago
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".β93Updated 11 months ago
- A tool for evaluating LLMsβ402Updated 9 months ago
- Mistral + Haystack: build RAG pipelines that rock π€β100Updated last year
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"β108Updated last year
- Sample notebooks and prompts for LLM evaluationβ120Updated 2 months ago
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.β106Updated this week
- This is the code for the paper "Self-contradictory Hallucinations of Large Language Models: Evaluation, Detection and Mitigation".β35Updated 10 months ago
- The official code of LM-Debugger, an interactive tool for inspection and intervention in transformer-based language models.β178Updated 2 years ago
- Interpretability for sequence generation models π πβ401Updated 3 months ago