stanford-crfm / EUAIActJune15Links
Stanford CRFM's initiative to assess potential compliance with the draft EU AI Act
β93Updated 2 years ago
Alternatives and similar repositories for EUAIActJune15
Users that are interested in EUAIActJune15 are comparing it to the libraries listed below
Sorting:
- Fiddler Auditor is a tool to evaluate language models.β188Updated last year
- π A curated list of papers & technical articles on AI Quality & Safetyβ192Updated 6 months ago
- The Foundation Model Transparency Indexβ83Updated last year
- β268Updated 9 months ago
- β256Updated 7 months ago
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.β108Updated this week
- Command Line Interface for Hugging Face Inference Endpointsβ66Updated last year
- AI Data Management & Evaluation Platformβ216Updated 2 years ago
- π€ Disaggregators: Curated data labelers for in-depth analysis.β67Updated 2 years ago
- π A curated list of resources dedicated to synthetic dataβ137Updated 3 years ago
- ReLM is a Regular Expression engine for Language Modelsβ106Updated 2 years ago
- β84Updated 2 years ago
- Framework for building and maintaining self-updating prompts for LLMsβ64Updated last year
- Mixing Language Models with Self-Verification and Meta-Verificationβ109Updated 10 months ago
- β100Updated last year
- β48Updated last year
- Make it easy to automatically and uniformly measure the behavior of many AI Systems.β26Updated last year
- Erasing concepts from neural representations with provable guaranteesβ238Updated 9 months ago
- An open-source compliance-centered evaluation framework for Generative AI modelsβ169Updated this week
- Resources related to EACL 2023 paper "SwitchPrompt: Learning Domain-Specific Gated Soft Prompts for Classification in Low-Resource Domainβ¦β52Updated 2 years ago
- π¦ Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data β¦β211Updated last week
- β78Updated last year
- codebase release for EMNLP2023 paper publicationβ19Updated last month
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"β107Updated 2 years ago
- β95Updated 10 months ago
- Credo AI Lens is a comprehensive assessment framework for AI systems. Lens standardizes model and data assessment, and acts as a central β¦β47Updated last year
- π Reference-Free automatic summarization evaluation with potential hallucination detectionβ102Updated last year
- The AI Incident Database seeks to identify, define, and catalog artificial intelligence incidents.β205Updated last week
- Check for data drift between two OpenAI multi-turn chat jsonl files.β38Updated last year
- Public blueprints for data use casesβ85Updated last month