mlcommons / modelgauge
Make it easy to automatically and uniformly measure the behavior of many AI Systems.
☆25Updated last week
Related projects: ⓘ
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆50Updated this week
- A mechanistic approach for understanding and detecting factual errors of large language models.☆38Updated 2 months ago
- Sparse and discrete interpretability tool for neural networks☆51Updated 7 months ago
- Investigating the generalization behavior of LM probes trained to predict truth labels: (1) from one annotator to another, and (2) from e…☆23Updated 3 months ago
- ☆47Updated 3 months ago
- Code for the ACL 2023 paper: "Rethinking the Role of Scale for In-Context Learning: An Interpretability-based Case Study at 66 Billion Sc…☆25Updated last year
- datasets from the paper "Towards Understanding Sycophancy in Language Models"☆59Updated 10 months ago
- Code and results accompanying the paper "Refusal in Language Models Is Mediated by a Single Direction".☆76Updated 3 weeks ago
- Adversarial Attacks on GPT-4 via Simple Random Search [Dec 2023]☆41Updated 4 months ago
- The Official Repository for "Bring Your Own Data! Self-Supervised Evaluation for Large Language Models"☆109Updated 11 months ago
- This repo contains code for the paper: "Can Foundation Models Help Us Achieve Perfect Secrecy?"☆24Updated last year
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆56Updated 3 months ago
- ☆91Updated last month
- ☆68Updated 3 weeks ago
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing System☆86Updated 3 months ago
- ☆23Updated last year
- WMDP is a LLM proxy benchmark for hazardous knowledge in bio, cyber, and chemical security. We also release code for RMU, an unlearning m…☆72Updated 4 months ago
- Scaling is a distributed training library and installable dependency designed to scale up neural networks, with a dedicated module for tr…☆38Updated 3 weeks ago
- Röttger et al. (2023): "XSTest: A Test Suite for Identifying Exaggerated Safety Behaviours in Large Language Models"☆55Updated 8 months ago
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.☆28Updated this week
- ☆38Updated 8 months ago
- ☆54Updated last week
- ☆66Updated last month
- Data for "Datamodels: Predicting Predictions with Training Data"☆87Updated last year
- ☆27Updated last year
- Steering vectors for transformer language models in Pytorch / Huggingface☆52Updated 2 months ago
- This is the repository holding code and data for "FrugalML: How to Use ML Prediction APIs More Accurately and Cheaply".☆30Updated 3 years ago
- Code for reproducing our paper "Not All Language Model Features Are Linear"☆57Updated last week
- ☆49Updated last year
- Get language models to generate responses in a specific format reliably. Open source implementation of Synchromesh: Reliable code generat…☆23Updated 6 months ago