kakaobrain / coyo-alignLinks
ALIGN trained on COYO-dataset
☆29Updated last year
Alternatives and similar repositories for coyo-align
Users that are interested in coyo-align are comparing it to the libraries listed below
Sorting:
- ViT trained on COYO-Labeled-300M dataset☆32Updated 2 years ago
- ☆65Updated 2 years ago
- Official implementation of "Active Image Indexing"☆59Updated 2 years ago
- ☆87Updated last year
- OCR-VQGAN, a discrete image encoder (tokenizer and detokenizer) for figure images in Paper2Fig100k dataset. Implementation of OCR Percept…☆81Updated 2 years ago
- FuseCap: Leveraging Large Language Models for Enriched Fused Image Captions☆55Updated last year
- [ECCV2024][ICCV2023] Official PyTorch implementation of SeiT++ and SeiT☆55Updated last year
- Un-*** 50 billions multimodality dataset☆22Updated 3 years ago
- [BMVC22] Official Implementation of ViCHA: "Efficient Vision-Language Pretraining with Visual Concepts and Hierarchical Alignment"☆55Updated 2 years ago
- Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training☆138Updated 2 years ago
- ☆46Updated last year
- The official PyTorch implementation for arXiv'23 paper 'LayoutDETR: Detection Transformer Is a Good Multimodal Layout Designer'☆100Updated 4 months ago
- ☆103Updated last year
- Official repository for the General Robust Image Task (GRIT) Benchmark☆54Updated 2 years ago
- Repository for the paper "Data Efficient Masked Language Modeling for Vision and Language".☆18Updated 4 years ago
- https://arxiv.org/abs/2209.15162☆52Updated 2 years ago
- Official Pytorch implementation of "CompoDiff: Versatile Composed Image Retrieval With Latent Diffusion" (TMLR 2024)☆87Updated 8 months ago
- Implementation of MaMMUT, a simple vision-encoder text-decoder architecture for multimodal tasks from Google, in Pytorch☆102Updated 2 years ago
- PyTorch code for MUST☆107Updated 5 months ago
- ☆53Updated 3 years ago
- Reproducible scaling laws for contrastive language-image learning (https://arxiv.org/abs/2212.07143)☆177Updated 3 months ago
- Code for experiments for "ConvNet vs Transformer, Supervised vs CLIP: Beyond ImageNet Accuracy"☆101Updated last year
- ECCV2024_Parrot Captions Teach CLIP to Spot Text☆65Updated last year
- Using pretrained encoder and language models to generate captions from multimedia inputs.☆97Updated 2 years ago
- codebase for the SIMAT dataset and evaluation☆38Updated 3 years ago
- ImageNet-12k subset of ImageNet-21k (fall11)☆21Updated 2 years ago
- ECCV2020 paper: Fashion Captioning: Towards Generating Accurate Descriptions with Semantic Rewards. Code and Data.☆85Updated 2 years ago
- Command-line tool for downloading and extending the RedCaps dataset.☆48Updated last year
- [ICML 2025] This is the official repository of our paper "What If We Recaption Billions of Web Images with LLaMA-3 ?"☆142Updated last year
- Patching open-vocabulary models by interpolating weights☆91Updated 2 years ago