socialfoundations / benchbenchLinks
BenchBench is a Python package to evaluate multi-task benchmarks.
☆16Updated last year
Alternatives and similar repositories for benchbench
Users that are interested in benchbench are comparing it to the libraries listed below
Sorting:
- DiWA: Diverse Weight Averaging for Out-of-Distribution Generalization☆31Updated 2 years ago
- Recycling diverse models☆45Updated 2 years ago
- ModelDiff: A Framework for Comparing Learning Algorithms☆59Updated last year
- ☆37Updated last year
- Code and results accompanying our paper titled RLSbench: Domain Adaptation under Relaxed Label Shift☆34Updated 2 years ago
- ☆34Updated 2 months ago
- ☆19Updated 3 years ago
- Data for "Datamodels: Predicting Predictions with Training Data"☆97Updated 2 years ago
- ☆18Updated 2 years ago
- ☆108Updated 2 years ago
- Model Patching: Closing the Subgroup Performance Gap with Data Augmentation☆42Updated 4 years ago
- ☆46Updated 2 years ago
- Code accompanying paper: Meta-Learning to Improve Pre-Training☆37Updated 3 years ago
- ☆27Updated 2 years ago
- Fast Axiomatic Attribution for Neural Networks (NeurIPS*2021)☆16Updated 2 years ago
- Latest Weight Averaging (NeurIPS HITY 2022)☆31Updated 2 years ago
- Code for the CVPR 2021 paper: Understanding Failures of Deep Networks via Robust Feature Extraction☆36Updated 3 years ago
- Code for paper "Can contrastive learning avoid shortcut solutions?" NeurIPS 2021.☆47Updated 3 years ago
- This repository contains the code of the distribution shift framework presented in A Fine-Grained Analysis on Distribution Shift (Wiles e…☆83Updated last month
- LISA for ICML 2022☆50Updated 2 years ago
- This repository holds code and other relevant files for the NeurIPS 2022 tutorial: Foundational Robustness of Foundation Models.☆71Updated 2 years ago
- ☆95Updated 2 years ago
- The official repository for "Intermediate Layers Matter in Momentum Contrastive Self Supervised Learning" paper.☆40Updated 3 years ago
- Framework code with wandb, checkpointing, logging, configs, experimental protocols. Useful for fine-tuning models or training from scratc…☆151Updated 2 years ago
- Distilling Model Failures as Directions in Latent Space☆47Updated 2 years ago
- Pytorch code for "Improving Self-Supervised Learning by Characterizing Idealized Representations"☆41Updated 2 years ago
- Code for the paper "Data Feedback Loops: Model-driven Amplification of Dataset Biases"☆16Updated 2 years ago
- ☆37Updated 3 years ago
- Official code for "In Search of Robust Measures of Generalization" (NeurIPS 2020)☆28Updated 4 years ago
- ☆108Updated last year