mlfoundations / imagenet-captionsLinks
Release of ImageNet-Captions
☆50Updated 2 years ago
Alternatives and similar repositories for imagenet-captions
Users that are interested in imagenet-captions are comparing it to the libraries listed below
Sorting:
- ☆51Updated 2 years ago
- PyTorch code for MUST☆108Updated 2 months ago
- Command-line tool for downloading and extending the RedCaps dataset.☆48Updated last year
- Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training☆137Updated 2 years ago
- ☆34Updated 2 years ago
- Patching open-vocabulary models by interpolating weights☆91Updated last year
- ☆46Updated last year
- This code provides a PyTorch implementation for OTTER (Optimal Transport distillation for Efficient zero-shot Recognition), as described …☆69Updated 3 years ago
- ☆120Updated 2 years ago
- DALL-Eval: Probing the Reasoning Skills and Social Biases of Text-to-Image Generation Models (ICCV 2023)☆140Updated last month
- MetaShift: A Dataset of Datasets for Evaluating Contextual Distribution Shifts and Training Conflicts (ICLR 2022)☆109Updated 2 years ago
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities☆78Updated 3 years ago
- https://arxiv.org/abs/2209.15162☆50Updated 2 years ago
- L-Verse: Bidirectional Generation Between Image and Text☆108Updated 3 months ago
- ☆104Updated last year
- [NeurIPS 2021] ORL: Unsupervised Object-Level Representation Learning from Scene Images☆58Updated 3 years ago
- [ECCV2022] New benchmark for evaluating pre-trained model; New supervised contrastive learning framework.☆108Updated last year
- Generate text captions for images from their embeddings.☆110Updated last year
- Code for the paper titled "CiT Curation in Training for Effective Vision-Language Data".☆78Updated 2 years ago
- A task-agnostic vision-language architecture as a step towards General Purpose Vision☆92Updated 4 years ago
- Using pretrained encoder and language models to generate captions from multimedia inputs.☆97Updated 2 years ago
- FuseCap: Leveraging Large Language Models for Enriched Fused Image Captions☆55Updated last year
- Official repository for the General Robust Image Task (GRIT) Benchmark☆54Updated 2 years ago
- Create generated datasets and train robust classifiers☆36Updated last year
- Code and Models for "GeneCIS A Benchmark for General Conditional Image Similarity"☆59Updated 2 years ago
- Repository for the paper "Data Efficient Masked Language Modeling for Vision and Language".☆18Updated 3 years ago
- VideoCC is a dataset containing (video-URL, caption) pairs for training video-text machine learning models. It is created using an automa…☆78Updated 2 years ago
- Compress conventional Vision-Language Pre-training data☆51Updated last year
- A PyTorch implementation of Mugs proposed by our paper "Mugs: A Multi-Granular Self-Supervised Learning Framework".☆83Updated last year
- Code release for "Improved baselines for vision-language pre-training"☆60Updated last year