Data-Provenance-Initiative / Data-Provenance-CollectionLinks
☆261Updated 10 months ago
Alternatives and similar repositories for Data-Provenance-Collection
Users that are interested in Data-Provenance-Collection are comparing it to the libraries listed below
Sorting:
- ☆270Updated last year
- What's In My Big Data (WIMBD) - a toolkit for analyzing large text datasets☆227Updated last year
- ☆100Updated last year
- Evaluating LLMs with fewer examples☆169Updated last year
- Website for hosting the Open Foundation Models Cheat Sheet.☆269Updated 9 months ago
- datasets from the paper "Towards Understanding Sycophancy in Language Models"☆102Updated 2 years ago
- ☆80Updated last year
- Code accompanying "How I learned to start worrying about prompt formatting".☆115Updated 8 months ago
- PAIR.withgoogle.com and friend's work on interpretability methods☆220Updated this week
- The Foundation Model Transparency Index☆85Updated last month
- The GitHub repo for Goal Driven Discovery of Distributional Differences via Language Descriptions☆72Updated 2 years ago
- Functional Benchmarks and the Reasoning Gap☆89Updated last year
- Keeping language models honest by directly eliciting knowledge encoded in their activations.☆217Updated last week
- [Data + code] ExpertQA : Expert-Curated Questions and Attributed Answers☆136Updated last year
- Steering vectors for transformer language models in Pytorch / Huggingface☆140Updated 11 months ago
- awesome synthetic (text) datasets☆321Updated last month
- Tools for understanding how transformer predictions are built layer-by-layer☆567Updated 6 months ago
- ☆112Updated 11 months ago
- ☆91Updated last month
- A toolkit for describing model features and intervening on those features to steer behavior.☆228Updated last month
- ☆152Updated 5 months ago
- Erasing concepts from neural representations with provable guarantees☆243Updated last year
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆107Updated 2 years ago
- NAACL 2024. Code & Dataset for "🌁 Bridging the Novice-Expert Gap via Models of Decision-Making: A Case Study on Remediating Math Mistake…☆45Updated last year
- Datasets collection and preprocessings framework for NLP extreme multitask learning☆192Updated 6 months ago
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆117Updated this week
- ☆304Updated 2 years ago
- Let's build better datasets, together!☆269Updated last year
- RuLES: a benchmark for evaluating rule-following in language models☆248Updated 11 months ago
- Repository for research in the field of Responsible NLP at Meta.☆205Updated this week