stanford-crfm / EUAIActJune15Links
Stanford CRFM's initiative to assess potential compliance with the draft EU AI Act
☆93Updated last year
Alternatives and similar repositories for EUAIActJune15
Users that are interested in EUAIActJune15 are comparing it to the libraries listed below
Sorting:
- Fiddler Auditor is a tool to evaluate language models.☆187Updated last year
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆192Updated 5 months ago
- ☆267Updated 7 months ago
- The Foundation Model Transparency Index☆82Updated last year
- Command Line Interface for Hugging Face Inference Endpoints☆66Updated last year
- AI Data Management & Evaluation Platform☆216Updated last year
- ☆248Updated 5 months ago
- 📖 A curated list of resources dedicated to synthetic data☆135Updated 3 years ago
- Mixing Language Models with Self-Verification and Meta-Verification☆110Updated 9 months ago
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆104Updated this week
- Framework for building and maintaining self-updating prompts for LLMs☆64Updated last year
- Find and fix bugs in natural language machine learning models using adaptive testing.☆185Updated last year
- ☆82Updated 2 years ago
- 🤗 Disaggregators: Curated data labelers for in-depth analysis.☆67Updated 2 years ago
- ReLM is a Regular Expression engine for Language Models☆106Updated 2 years ago
- Topic modeling helpers using managed language models from Cohere. Name text clusters using large GPT models.☆223Updated 2 years ago
- This is an open-source tool to assess and improve the trustworthiness of AI systems.☆96Updated 2 weeks ago
- git extension for {collaborative, communal, continual} model development☆216Updated 10 months ago
- ☆80Updated last year
- ☆96Updated last year
- ☆337Updated last year
- Library for iPython notebooks for evaluating factuality.☆51Updated 2 years ago
- Make it easy to automatically and uniformly measure the behavior of many AI Systems.☆26Updated 11 months ago
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆114Updated last month
- ☆78Updated last year
- ☆169Updated 2 weeks ago
- Check for data drift between two OpenAI multi-turn chat jsonl files.☆37Updated last year
- Domain Adapted Language Modeling Toolkit - E2E RAG☆329Updated 10 months ago
- Leverage your LangChain trace data for fine tuning☆46Updated last year
- Doing simple retrieval from LLM models at various context lengths to measure accuracy☆103Updated last year