stanford-crfm / fmtiLinks
The Foundation Model Transparency Index
☆82Updated last year
Alternatives and similar repositories for fmti
Users that are interested in fmti are comparing it to the libraries listed below
Sorting:
- ☆249Updated 5 months ago
- ☆267Updated 7 months ago
- Stanford CRFM's initiative to assess potential compliance with the draft EU AI Act☆93Updated last year
- codebase release for EMNLP2023 paper publication☆19Updated this week
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆208Updated this week
- Functional Benchmarks and the Reasoning Gap☆88Updated 11 months ago
- AI Data Management & Evaluation Platform☆216Updated last year
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆192Updated 5 months ago
- This is the reproduction repository for my 🤗 Hugging Face blog post on synthetic data☆68Updated last year
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.☆168Updated this week
- ☆142Updated 2 weeks ago
- Evaluating LLMs with CommonGen-Lite☆91Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆110Updated 9 months ago
- ☆69Updated last year
- ReLM is a Regular Expression engine for Language Models☆106Updated 2 years ago
- 🤗 Disaggregators: Curated data labelers for in-depth analysis.☆67Updated 2 years ago
- This is an open-source tool to assess and improve the trustworthiness of AI systems.☆96Updated 2 weeks ago
- Open Implementations of LLM Analyses☆107Updated 11 months ago
- ☆43Updated last year
- This repository's goal is to precompile all past presentations of the Huggingface reading group☆48Updated last year
- ☆81Updated 2 weeks ago
- ☆163Updated last year
- Website for hosting the Open Foundation Models Cheat Sheet.☆268Updated 4 months ago
- CiteME is a benchmark designed to test the abilities of language models in finding papers that are cited in scientific texts.☆48Updated 10 months ago
- A repository containing the code for translating popular LLM benchmarks to German.☆29Updated 2 years ago
- Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning☆46Updated last year
- ☆96Updated last year
- Evaluating LLMs with fewer examples☆161Updated last year
- GraphER: A Structure-aware Text-to-Graph Model for Entity and Relation Extraction☆80Updated last year
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆104Updated this week