mlfoundations / model-soups
Model soups: averaging weights of multiple fine-tuned models improves accuracy without increasing inference time
☆437Updated 6 months ago
Alternatives and similar repositories for model-soups:
Users that are interested in model-soups are comparing it to the libraries listed below
- Code release for "Dropout Reduces Underfitting"☆311Updated last year
- Robust fine-tuning of zero-shot models☆665Updated 2 years ago
- A framework for merging models solving different tasks with different initializations into one multi-task model without any additional tr…☆290Updated last year
- Official code for "TOAST: Transfer Learning via Attention Steering"☆186Updated last year
- ☆198Updated last year
- DataComp: In search of the next generation of multimodal datasets☆674Updated last year
- Official code for our CVPR'22 paper “Vision Transformer Slimming: Multi-Dimension Searching in Continuous Optimization Space”☆247Updated last year
- Compare neural networks by their feature similarity☆349Updated last year
- Learning from synthetic data - code and models☆307Updated last year
- Masked Siamese Networks for Label-Efficient Learning (https://arxiv.org/abs/2204.07141)☆453Updated 2 years ago
- PyTorch code for hierarchical k-means -- a data curation method for self-supervised learning☆141Updated 7 months ago
- [NeurIPS 2023] This repository includes the official implementation of our paper "An Inverse Scaling Law for CLIP Training"☆309Updated 7 months ago
- A collection of parameter-efficient transfer learning papers focusing on computer vision and multimodal domains.☆396Updated 4 months ago
- ☆600Updated last week
- Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch☆299Updated 7 months ago
- FFCV-SSL Fast Forward Computer Vision for Self-Supervised Learning.☆203Updated last year
- ☆486Updated 2 months ago
- Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorch☆257Updated 9 months ago
- (ICLR 2022 Spotlight) Official PyTorch implementation of "How Do Vision Transformers Work?"☆808Updated 2 years ago
- Supervision Exists Everywhere: A Data Efficient Contrastive Language-Image Pre-training Paradigm☆643Updated 2 years ago
- Exploring Visual Prompts for Adapting Large-Scale Models☆271Updated 2 years ago
- EsViT: Efficient self-supervised Vision Transformers☆409Updated last year
- ☆181Updated last year
- CLIP-like model evaluation☆656Updated 5 months ago
- When do we not need larger vision models?☆357Updated last month
- Probing the representations of Vision Transformers.☆319Updated 2 years ago
- Learning to Prompt (L2P) for Continual Learning @ CVPR22 and DualPrompt: Complementary Prompting for Rehearsal-free Continual Learning @ …☆424Updated 5 months ago
- Reproducible scaling laws for contrastive language-image learning (https://arxiv.org/abs/2212.07143)☆158Updated last year
- Benchmark your model on out-of-distribution datasets with carefully collected human comparison data (NeurIPS 2021 Oral)☆339Updated 5 months ago
- This repository contains the implementation for the paper "EMP-SSL: Towards Self-Supervised Learning in One Training Epoch."☆224Updated last year