Code and models for the paper "The effectiveness of MAE pre-pretraining for billion-scale pretraining" https://arxiv.org/abs/2303.13496
☆92Feb 19, 2026Updated last week
Alternatives and similar repositories for maws
Users that are interested in maws are comparing it to the libraries listed below
Sorting:
- PyTorch code and pretrained weights for the UNIC models.☆44Aug 29, 2024Updated last year
- Checkpointable dataset utilities for foundation model training☆32Jan 29, 2024Updated 2 years ago
- [IEEE T-BIOM] FaceXBench: Evaluating Multimodal LLMs on Face Understanding☆20Jan 15, 2026Updated last month
- ☆17Aug 7, 2024Updated last year
- Official Pytorch codebase for Open-Vocabulary Instance Segmentation without Manual Mask Annotations [CVPR 2023]☆52Oct 26, 2025Updated 4 months ago
- MMPD Dataset from ECCV'2024 "When Pedestrian Detection Meets Multi-Modal Learning: Generalist Model and Benchmark Dataset"☆21Jul 15, 2024Updated last year
- Official Repo for PosSAM: Panoptic Open-vocabulary Segment Anything☆70Apr 7, 2024Updated last year
- Official Pytorch Implementation of Self-emerging Token Labeling☆35Mar 27, 2024Updated last year
- The official repo for LIFT: Language-Image Alignment with Fixed Text Encoders☆42Jun 10, 2025Updated 8 months ago
- ☆19Dec 6, 2023Updated 2 years ago
- HT-Step is a large-scale article grounding dataset of temporal step annotations on how-to videos☆24Mar 20, 2024Updated last year
- Pytorch code for paper From CLIP to DINO: Visual Encoders Shout in Multi-modal Large Language Models☆207Jan 8, 2025Updated last year
- Omnivore: A Single Model for Many Visual Modalities☆571Nov 12, 2022Updated 3 years ago
- 1.5−3.0× lossless training or pre-training speedup. An off-the-shelf, easy-to-implement algorithm for the efficient training of foundatio…☆226Aug 23, 2024Updated last year
- [NeurIPS 2023] This repo contains the code for our paper Convolutions Die Hard: Open-Vocabulary Segmentation with Single Frozen Convoluti…☆337Feb 5, 2024Updated 2 years ago
- ☆12Aug 19, 2023Updated 2 years ago
- Implementation of the spotlight: a method for discovering systematic errors in deep learning models☆11Oct 5, 2021Updated 4 years ago
- [ICLR'25] Do Egocentric Video-Language Models Truly Understand Hand-Object Interactions?☆12Apr 11, 2025Updated 10 months ago
- A Novel Semantic Segmentation Network using Enhanced Boundaries in Cluttered Scenes (WACV 2025)☆11Aug 11, 2025Updated 6 months ago
- NeurIPS 2025 Spotlight; ICLR2024 Spotlight; CVPR 2024; EMNLP 2024☆1,815Nov 27, 2025Updated 3 months ago
- High-performance Image Tokenizers for VAR and AR☆303Apr 25, 2025Updated 10 months ago
- DataComp: In search of the next generation of multimodal datasets☆772Apr 28, 2025Updated 10 months ago
- Hiera: A fast, powerful, and simple hierarchical vision transformer.☆1,055Mar 2, 2024Updated 2 years ago
- Sambor: Boosting Segment Anything Model Towards Open-Vocabulary Learning☆32Dec 7, 2023Updated 2 years ago
- Original code base for On Pretraining Data Diversity for Self-Supervised Learning☆14Dec 30, 2024Updated last year
- Code for the paper "Stack Attention: Improving the Ability of Transformers to Model Hierarchical Patterns"☆18Mar 15, 2024Updated last year
- "Roll with the Punches: Expansion and Shrinkage of Soft Label Selection for Semi-supervised Fine-Grained Learning" by Yue Duan (AAAI 2024…☆13Nov 20, 2025Updated 3 months ago
- recipe for training fully-featured self supervised image jepa models☆12Jun 4, 2025Updated 8 months ago
- Official Implementation of the CrossMAE paper: Rethinking Patch Dependence for Masked Autoencoders☆132Apr 10, 2025Updated 10 months ago
- [ICLR 2026] Official implementation for What matters for Representation Alignment: Global Information or Spatial Structure?☆221Dec 15, 2025Updated 2 months ago
- Densely Captioned Images (DCI) dataset repository.☆197Jul 1, 2024Updated last year
- Training code for CLIP-FlanT5☆30Jul 29, 2024Updated last year
- ☆32Jul 29, 2024Updated last year
- ☆548Nov 7, 2024Updated last year
- IMProv: Inpainting-based Multimodal Prompting for Computer Vision Tasks☆58Sep 26, 2024Updated last year
- [ICLR 2025] AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark☆138Jun 4, 2025Updated 8 months ago
- Official Repo for the paper: VCR: Visual Caption Restoration. Check arxiv.org/pdf/2406.06462 for details.☆32Feb 26, 2025Updated last year
- EVE Series: Encoder-Free Vision-Language Models from BAAI☆368Jul 24, 2025Updated 7 months ago
- When do we not need larger vision models?☆413Feb 8, 2025Updated last year