stanford-crfm / EUAIActJune15
Stanford CRFM's initiative to assess potential compliance with the draft EU AI Act
☆94Updated last year
Alternatives and similar repositories for EUAIActJune15:
Users that are interested in EUAIActJune15 are comparing it to the libraries listed below
- The Foundation Model Transparency Index☆78Updated 11 months ago
- Fiddler Auditor is a tool to evaluate language models.☆179Updated last year
- 🤗 Disaggregators: Curated data labelers for in-depth analysis.☆65Updated 2 years ago
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆88Updated this week
- ☆230Updated last month
- Make it easy to automatically and uniformly measure the behavior of many AI Systems.☆27Updated 6 months ago
- A comprehensive guide to LLM evaluation methods designed to assist in identifying the most suitable evaluation techniques for various use…☆114Updated last week
- ☆264Updated 3 months ago
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆178Updated last week
- TalkToModel gives anyone with the powers of XAI through natural language conversations 💬!☆120Updated last year
- Command Line Interface for Hugging Face Inference Endpoints☆66Updated last year
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆108Updated last year
- Framework for building and maintaining self-updating prompts for LLMs☆61Updated 10 months ago
- ReLM is a Regular Expression engine for Language Models☆103Updated last year
- Functional Benchmarks and the Reasoning Gap☆85Updated 6 months ago
- ☆51Updated 10 months ago
- ☆78Updated 2 years ago
- ☆77Updated 10 months ago
- Credo AI Lens is a comprehensive assessment framework for AI systems. Lens standardizes model and data assessment, and acts as a central …☆47Updated 10 months ago
- ☆24Updated last year
- ☆67Updated 5 months ago
- ☆68Updated last year
- Red-Teaming Language Models with DSPy☆183Updated 2 months ago
- A collection of machine learning model cards and datasheets.☆75Updated 10 months ago
- 🧠Starter templates for doing interpretability research☆70Updated last year
- ☆93Updated 11 months ago
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.☆124Updated this week
- Collection of evals for Inspect AI☆115Updated this week
- Track the progress of LLM context utilisation☆54Updated last week
- git extension for {collaborative, communal, continual} model development☆211Updated 5 months ago