stanford-crfm / ecosystem-graphsLinks
☆268Updated 9 months ago
Alternatives and similar repositories for ecosystem-graphs
Users that are interested in ecosystem-graphs are comparing it to the libraries listed below
Sorting:
- ☆256Updated 7 months ago
- Website for hosting the Open Foundation Models Cheat Sheet.☆268Updated 6 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆256Updated 2 years ago
- Build, evaluate, understand, and fix LLM-based apps☆491Updated last year
- RuLES: a benchmark for evaluating rule-following in language models☆239Updated 8 months ago
- A joint community effort to create one central leaderboard for LLMs.☆305Updated last year
- ☆315Updated last year
- ☆297Updated 2 years ago
- Scaling Data-Constrained Language Models☆341Updated 4 months ago
- ☆149Updated last year
- Project 2 (Building Large Language Models) for Stanford CS324: Understanding and Developing Large Language Models (Winter 2022)☆105Updated 2 years ago
- Evaluating LLMs with fewer examples☆168Updated last year
- The official evaluation suite and dynamic data release for MixEval.☆252Updated last year
- [ICLR 2024 Spotlight] FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets☆217Updated last year
- PaL: Program-Aided Language Models (ICML 2023)☆515Updated 2 years ago
- Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning☆46Updated last year
- ArXiv Digest and Personalized Recommendations using Large Language Models☆382Updated last year
- ☆100Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆109Updated 11 months ago
- 📚 A curated list of papers & technical articles on AI Quality & Safety☆193Updated 7 months ago
- ☆43Updated last year
- Repo for the Belebele dataset, a massively multilingual reading comprehension dataset.☆336Updated 11 months ago
- awesome synthetic (text) datasets☆305Updated this week
- data cleaning and curation for unstructured text☆329Updated last year
- [ICLR 2024 & NeurIPS 2023 WS] An Evaluator LM that is open-source, offers reproducible evaluation, and inexpensive to use. Specifically d…☆307Updated 2 years ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆312Updated last year
- Evaluation suite for LLMs☆365Updated 4 months ago
- Code and data for "Measuring and Narrowing the Compositionality Gap in Language Models"☆323Updated last year
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆224Updated 2 months ago
- Benchmarking long-form factuality in large language models. Original code for our paper "Long-form factuality in large language models".☆652Updated 3 months ago