Data-Provenance-Initiative / Data-Provenance-Collection
☆206Updated this week
Alternatives and similar repositories for Data-Provenance-Collection:
Users that are interested in Data-Provenance-Collection are comparing it to the libraries listed below
- awesome synthetic (text) datasets☆256Updated 3 months ago
- What's In My Big Data (WIMBD) - a toolkit for analyzing large text datasets☆204Updated 2 months ago
- Website for hosting the Open Foundation Models Cheat Sheet.☆263Updated 7 months ago
- ☆489Updated 2 months ago
- Evaluating LLMs with fewer examples☆141Updated 9 months ago
- Pretraining Efficiently on S2ORC!☆149Updated 3 months ago
- datasets from the paper "Towards Understanding Sycophancy in Language Models"☆67Updated last year
- Extract full next-token probabilities via language model APIs☆229Updated 11 months ago
- Let's build better datasets, together!☆250Updated last month
- ☆259Updated this week
- Automatic Evals for LLMs☆128Updated this week
- Manage scalable open LLM inference endpoints in Slurm clusters☆249Updated 6 months ago
- Functional Benchmarks and the Reasoning Gap☆82Updated 3 months ago
- RuLES: a benchmark for evaluating rule-following in language models☆214Updated last week
- Datasets collection and preprocessings framework for NLP extreme multitask learning☆173Updated 3 weeks ago
- [Data + code] ExpertQA : Expert-Curated Questions and Attributed Answers☆124Updated 10 months ago
- [ICLR 2024 Spotlight] FLASK: Fine-grained Language Model Evaluation based on Alignment Skill Sets☆214Updated last year
- This is the reproduction repository for my 🤗 Hugging Face blog post on synthetic data☆63Updated 11 months ago
- Code accompanying "How I learned to start worrying about prompt formatting".☆100Updated 3 months ago
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆108Updated last year
- Benchmarking LLMs with Challenging Tasks from Real Users☆208Updated 2 months ago
- ☆139Updated this week
- This project studies the performance and robustness of language models and task-adaptation methods.☆142Updated 8 months ago
- The official evaluation suite and dynamic data release for MixEval.☆233Updated 2 months ago
- Repository for research in the field of Responsible NLP at Meta.☆192Updated 2 months ago
- Code for the paper "Fishing for Magikarp"☆141Updated 2 weeks ago
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆75Updated this week
- ☆117Updated last week
- ☆100Updated 8 months ago
- Code and results accompanying the paper "Refusal in Language Models Is Mediated by a Single Direction".☆163Updated 3 months ago