PengtaoJiang / Segment-Anything-CLIPView external linksLinks
Connecting segment-anything's output masks with the CLIP model; Awesome-Segment-Anything-Works
☆201Oct 13, 2024Updated last year
Alternatives and similar repositories for Segment-Anything-CLIP
Users that are interested in Segment-Anything-CLIP are comparing it to the libraries listed below
Sorting:
- Experiment on combining CLIP with SAM to do open-vocabulary image segmentation.☆385Apr 5, 2023Updated 2 years ago
- Segment Anything combined with CLIP☆349Feb 19, 2024Updated last year
- Referring Image Segmentation Benchmarking with Segment Anything Model (SAM)☆38Apr 7, 2023Updated 2 years ago
- Recent weakly supervised semantic segmentation paper☆377Aug 26, 2025Updated 5 months ago
- Official implement of CVPR2023 ZegCLIP: Towards Adapting CLIP for Zero-shot Semantic Segmentation☆259Aug 27, 2023Updated 2 years ago
- Combining Segment Anything (SAM) with Grounded DINO for zero-shot object detection and CLIPSeg for zero-shot segmentation☆434May 3, 2024Updated last year
- [CVPR 2023] CLIP is Also an Efficient Segmenter: A Text-Driven Approach for Weakly Supervised Semantic Segmentation☆210Sep 16, 2024Updated last year
- (TPAMI 2024) A Survey on Open Vocabulary Learning☆986Dec 24, 2025Updated last month
- [ICLR2024 Spotlight] Code Release of CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense Prediction☆201Feb 5, 2024Updated 2 years ago
- [ECCV2024] This is an official implementation for "PSALM: Pixelwise SegmentAtion with Large Multi-Modal Model"☆269Dec 30, 2024Updated last year
- This method uses Segment Anything and CLIP to ground and count any object that matches a custom text prompt, without requiring any point …☆176Apr 22, 2023Updated 2 years ago
- [ICLR'24 & IJCV‘25] Matcher: Segment Anything with One Shot Using All-Purpose Feature Matching☆547Dec 3, 2025Updated 2 months ago
- [CVPR 2024 🔥] Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses tha…☆945Aug 5, 2025Updated 6 months ago
- [ECCV 2024] Official implementation of the paper "Semantic-SAM: Segment and Recognize Anything at Any Granularity"☆2,808Jul 10, 2025Updated 7 months ago
- Official Repo for PosSAM: Panoptic Open-vocabulary Segment Anything☆70Apr 7, 2024Updated last year
- A curated publication list on open vocabulary semantic segmentation and related area (e.g. zero-shot semantic segmentation) resources..☆828Jan 20, 2026Updated 3 weeks ago
- Official PyTorch implementation of "Extract Free Dense Labels from CLIP" (ECCV 22 Oral)☆470Sep 19, 2022Updated 3 years ago
- [CVPR 2023] This is the official PyTorch implementation for "Dynamic Focus-aware Positional Queries for Semantic Segmentation".☆61Mar 4, 2023Updated 2 years ago
- This is the official PyTorch implementation of the paper Open-Vocabulary Semantic Segmentation with Mask-adapted CLIP.☆752Oct 17, 2023Updated 2 years ago
- Open-vocabulary Semantic Segmentation☆371Oct 16, 2024Updated last year
- [ECCV 2024] The official code of paper "Open-Vocabulary SAM".☆1,028Aug 4, 2025Updated 6 months ago
- [ICCV 2023] CTVIS: Consistent Training for Online Video Instance Segmentation☆80Oct 15, 2023Updated 2 years ago
- The PyTorch Code for our ICCV 2019 paper "Integral Object Mining via Online Attention Accumulation"☆57Dec 29, 2021Updated 4 years ago
- 🔥 [CVPR 2024] Official implementation of "See, Say, and Segment: Teaching LMMs to Overcome False Premises (SESAME)"☆47Jun 16, 2024Updated last year
- Automated dense category annotation engine that serves as the initial semantic labeling for the Segment Anything dataset (SA-1B).☆2,297Jun 7, 2023Updated 2 years ago
- ☆22Jun 30, 2023Updated 2 years ago
- Fine-tune SAM (Segment Anything Model) for computer vision tasks such as semantic segmentation, matting, detection ... in specific scena…☆860Aug 5, 2023Updated 2 years ago
- [ICCV 2023] Official implementation of the paper "A Simple Framework for Open-Vocabulary Segmentation and Detection"☆748Jan 22, 2024Updated 2 years ago
- Codes for ICML 2023 Learning Dynamic Query Combinations for Transformer-based Object Detection and Segmentation☆37Sep 12, 2023Updated 2 years ago
- [CVPR 2023] Official code for "Zero-shot Referring Image Segmentation with Global-Local Context Features"☆128Mar 17, 2025Updated 10 months ago
- GroundVLP: Harnessing Zero-shot Visual Grounding from Vision-Language Pre-training and Open-Vocabulary Object Detection (AAAI 2024)☆72Jan 2, 2024Updated 2 years ago
- [ECCV 2024] Tokenize Anything via Prompting☆603Dec 11, 2024Updated last year
- A collection of project, papers, and source code for Meta AI's Segment Anything Model (SAM) and related studies.☆370Dec 8, 2024Updated last year
- [T-PAMI-2024] Transformer-Based Visual Segmentation: A Survey☆759Aug 25, 2024Updated last year
- Using Segment-Anything and CLIP to generate pixel-aligned semantic features.☆40Apr 27, 2023Updated 2 years ago
- LLM-Seg: Bridging Image Segmentation and Large Language Model Reasoning☆195Apr 16, 2024Updated last year
- [ICCV 2023] OnlineRefer: A Simple Online Baseline for Referring Video Object Segmentation☆57Oct 7, 2023Updated 2 years ago
- ☆37Oct 21, 2022Updated 3 years ago
- Code release for "BoxVIS: Video Instance Segmentation with Box Annotation"☆12Dec 22, 2023Updated 2 years ago