dfan / websslLinks
Code for Scaling Language-Free Visual Representation Learning (WebSSL)
☆245Updated 8 months ago
Alternatives and similar repositories for webssl
Users that are interested in webssl are comparing it to the libraries listed below
Sorting:
- Code for "Scaling Language-Free Visual Representation Learning" paper (Web-SSL).☆194Updated 8 months ago
- [CVPR 2025] FLAIR: VLM with Fine-grained Language-informed Image Representations☆129Updated 4 months ago
- [NeurIPS 2024] Classification Done Right for Vision-Language Pre-Training☆225Updated 9 months ago
- [ICLR'25] Reconstructive Visual Instruction Tuning☆133Updated 9 months ago
- Official implementation of 'CLIP-DINOiser: Teaching CLIP a few DINO tricks' paper.☆270Updated last year
- DenseFusion-1M: Merging Vision Experts for Comprehensive Multimodal Perception☆158Updated last year
- Large-Vocabulary Video Instance Segmentation dataset☆95Updated last year
- [ECCV2024 Oral🔥] Official Implementation of "GiT: Towards Generalist Vision Transformer through Universal Language Interface"☆358Updated last year
- Open source implementation of "Vision Transformers Need Registers"☆206Updated this week
- [ICLR 2025] Diffusion Feedback Helps CLIP See Better☆299Updated 11 months ago
- [ECCV24] VISA: Reasoning Video Object Segmentation via Large Language Model☆203Updated last year
- [CVPR'24] The repository provides code for running inference and training for "Segment and Caption Anything" (SCA) , links for downloadin…☆231Updated last year
- [ECCV2024] PartGLEE: A Foundation Model for Recognizing and Parsing Any Objects☆57Updated last year
- PyTorch implementation of NEPA☆262Updated 3 weeks ago
- [ECCV 2024] Official PyTorch implementation of DreamLIP: Language-Image Pre-training with Long Captions☆136Updated 8 months ago
- When do we not need larger vision models?☆413Updated 11 months ago
- Official implementation of SCLIP: Rethinking Self-Attention for Dense Vision-Language Inference☆180Updated last year
- [CVPR 2025 🔥]A Large Multimodal Model for Pixel-Level Visual Grounding in Videos☆94Updated 9 months ago
- EVE Series: Encoder-Free Vision-Language Models from BAAI☆363Updated 5 months ago
- Densely Captioned Images (DCI) dataset repository.☆195Updated last year
- official implementation of "Interpreting CLIP's Image Representation via Text-Based Decomposition"☆233Updated 7 months ago
- ☆356Updated last year
- [CVPR24] Official Implementation of GEM (Grounding Everything Module)☆135Updated 9 months ago
- Code and dataset link for "DenseWorld-1M: Towards Detailed Dense Grounded Caption in the Real World"☆120Updated 3 months ago
- [NeurlPS 2024] One Token to Seg Them All: Language Instructed Reasoning Segmentation in Videos☆143Updated last year
- [COLM'25] Official implementation of the Law of Vision Representation in MLLMs☆172Updated 3 months ago
- [ICLR2024 Spotlight] Code Release of CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction☆201Updated last year
- This repo contains the official implementation of ICLR 2024 paper "Is ImageNet worth 1 video? Learning strong image encoders from 1 long …☆95Updated last year
- An open source implementation of CLIP (With TULIP Support)☆165Updated 8 months ago
- [CVPR2024] GSVA: Generalized Segmentation via Multimodal Large Language Models☆158Updated last year