stanford-crfm / EUAIActJune15Links
Stanford CRFM's initiative to assess potential compliance with the draft EU AI Act
☆93Updated 2 years ago
Alternatives and similar repositories for EUAIActJune15
Users that are interested in EUAIActJune15 are comparing it to the libraries listed below
Sorting:
- Fiddler Auditor is a tool to evaluate language models.☆188Updated last year
- The Foundation Model Transparency Index☆83Updated this week
- ☆270Updated 10 months ago
- Framework for building and maintaining self-updating prompts for LLMs☆64Updated last year
- 📖 A curated list of resources dedicated to synthetic data☆140Updated 3 years ago
- Command Line Interface for Hugging Face Inference Endpoints☆66Updated last year
- AI Data Management & Evaluation Platform☆216Updated 2 years ago
- ☆258Updated 8 months ago
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆194Updated 7 months ago
- 🤗 Disaggregators: Curated data labelers for in-depth analysis.☆67Updated 2 years ago
- ☆84Updated 2 years ago
- ☆339Updated last year
- ☆172Updated 3 weeks ago
- Check for data drift between two OpenAI multi-turn chat jsonl files.☆39Updated last year
- Introduction to Data-Centric AI, MIT IAP 2024 🤖☆103Updated 5 months ago
- 📚 Datasets and models for instruction-tuning☆238Updated 2 years ago
- This is an open-source tool to assess and improve the trustworthiness of AI systems.☆98Updated last week
- A suite of auto-regressive and Seq2Seq (sequence-to-sequence) transformer models for tabular and relational synthetic data generation.☆242Updated 4 months ago
- An open-source compliance-centered evaluation framework for Generative AI models☆174Updated this week
- Mixing Language Models with Self-Verification and Meta-Verification☆110Updated last year
- ☆29Updated 2 years ago
- Completion After Prompt Probability. Make your LLM make a choice☆82Updated last year
- Topic modeling helpers using managed language models from Cohere. Name text clusters using large GPT models.☆222Updated 2 years ago
- ☆78Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆51Updated last year
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆116Updated 4 months ago
- Leverage your LangChain trace data for fine tuning☆46Updated last year
- ☆23Updated 2 years ago
- Mistral + Haystack: build RAG pipelines that rock 🤘☆106Updated last year
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆107Updated 2 years ago