Giskard-AI / awesome-ai-safetyLinks
📚 A curated list of papers & technical articles on AI Quality & Safety
☆199Updated 9 months ago
Alternatives and similar repositories for awesome-ai-safety
Users that are interested in awesome-ai-safety are comparing it to the libraries listed below
Sorting:
- Fiddler Auditor is a tool to evaluate language models.☆188Updated last year
- ☆271Updated 11 months ago
- An open-source compliance-centered evaluation framework for Generative AI models☆178Updated 3 weeks ago
- 📖 A curated list of resources dedicated to synthetic data☆140Updated 3 years ago
- Red-Teaming Language Models with DSPy☆250Updated 11 months ago
- 📚 Datasets and models for instruction-tuning☆238Updated 2 years ago
- Stanford CRFM's initiative to assess potential compliance with the draft EU AI Act☆93Updated 2 years ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆448Updated last year
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆212Updated last week
- ☆261Updated 9 months ago
- Mixing Language Models with Self-Verification and Meta-Verification☆111Updated last year
- Moonshot - A simple and modular tool to evaluate and red-team any LLM application.☆302Updated last week
- A curated list of awesome publications and researchers on prompting framework updated and maintained by The Intelligent System Security (…☆85Updated last year
- A tool for evaluating LLMs☆427Updated last year
- The Foundation Model Transparency Index☆85Updated last month
- data cleaning and curation for unstructured text☆328Updated last year
- Toolkit for attaching, training, saving and loading of new heads for transformer models☆294Updated 10 months ago
- ☆42Updated last year
- Accompanying code and SEP dataset for the "Can LLMs Separate Instructions From Data? And What Do We Even Mean By That?" paper.☆58Updated 10 months ago
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆118Updated last week
- Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning☆46Updated 2 years ago
- A curated list of awesome synthetic data tools (open source and commercial).☆231Updated 2 years ago
- RAGElo is a set of tools that helps you selecting the best RAG-based LLM agents by using an Elo ranker☆126Updated 2 months ago
- ☆339Updated 2 years ago
- The fastest Trust Layer for AI Agents☆148Updated 7 months ago
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆116Updated 5 months ago
- AI Verify☆40Updated last week
- A collection of the the best ML and AI news every week (research, news, resources)☆173Updated 5 months ago
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".☆124Updated last year
- A repository of Language Model Vulnerabilities and Exposures (LVEs).☆112Updated last year