Giskard-AI / awesome-ai-safetyLinks
📚 A curated list of papers & technical articles on AI Quality & Safety
☆193Updated 5 months ago
Alternatives and similar repositories for awesome-ai-safety
Users that are interested in awesome-ai-safety are comparing it to the libraries listed below
Sorting:
- Fiddler Auditor is a tool to evaluate language models.☆188Updated last year
- ☆267Updated 8 months ago
- AI Verify☆33Updated 2 weeks ago
- 📚 Datasets and models for instruction-tuning☆239Updated 2 years ago
- Stanford CRFM's initiative to assess potential compliance with the draft EU AI Act☆93Updated last year
- Red-Teaming Language Models with DSPy☆216Updated 7 months ago
- PromptInject is a framework that assembles prompts in a modular fashion to provide a quantitative analysis of the robustness of LLMs to a…☆422Updated last year
- The Foundation Model Transparency Index☆83Updated last year
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆115Updated 2 months ago
- An open-source compliance-centered evaluation framework for Generative AI models☆165Updated last week
- A tool for evaluating LLMs☆423Updated last year
- ☆253Updated 6 months ago
- ☆43Updated last year
- Reward Model framework for LLM RLHF☆61Updated 2 years ago
- Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning☆46Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆110Updated 9 months ago
- data cleaning and curation for unstructured text☆328Updated last year
- Continuous Integration for LLM powered applications☆252Updated 2 years ago
- Open Implementations of LLM Analyses☆107Updated last year
- Framework for LLM evaluation, guardrails and security☆113Updated last year
- Weekly visualization report of Open LLM model performance based on 4 metrics.☆86Updated last year
- A curated list of awesome academic research, books, code of ethics, data sets, institutes, maturity models, newsletters, principles, podc…☆85Updated this week
- Creating the tools and data sets necessary to evaluate vulnerabilities in LLMs.☆26Updated 6 months ago
- Moonshot - A simple and modular tool to evaluate and red-team any LLM application.☆275Updated last month
- Deliver safe & effective language models☆543Updated last week
- ☆336Updated last year
- A curated list of awesome publications and researchers on prompting framework updated and maintained by The Intelligent System Security (…☆84Updated 8 months ago
- A collection of the the best ML and AI news every week (research, news, resources)☆170Updated 2 months ago
- Curation of prompts that are known to be adversarial to large language models☆186Updated 2 years ago
- 📖 A curated list of resources dedicated to synthetic data☆136Updated 3 years ago