awaisrauf / Awesome-CV-Foundational-ModelsLinks
☆524Updated 8 months ago
Alternatives and similar repositories for Awesome-CV-Foundational-Models
Users that are interested in Awesome-CV-Foundational-Models are comparing it to the libraries listed below
Sorting:
- This repo lists relevant papers summarized in our survey paper: A Systematic Survey of Prompt Engineering on Vision-Language Foundation …☆476Updated 4 months ago
- [Pattern Recognition 25] CLIP Surgery for Better Explainability with Enhancement in Open-Vocabulary Tasks☆437Updated 5 months ago
- A collection of papers on the topic of ``Computer Vision in the Wild (CVinW)''☆1,322Updated last year
- Low rank adaptation for Vision Transformer☆418Updated last year
- Official Open Source code for "Scaling Language-Image Pre-training via Masking"☆426Updated 2 years ago
- Holds code for our CVPR'23 tutorial: All Things ViTs: Understanding and Interpreting Attention in Vision.☆194Updated 2 years ago
- A curated list of papers, datasets and resources pertaining to open vocabulary object detection.☆341Updated 2 months ago
- A curated list of foundation models for vision and language tasks☆1,066Updated last month
- A curated list of prompt-based paper in computer vision and vision-language learning.☆921Updated last year
- A curated publication list on open vocabulary semantic segmentation and related area (e.g. zero-shot semantic segmentation) resources..☆683Updated this week
- A collection of parameter-efficient transfer learning papers focusing on computer vision and multimodal domains.☆402Updated 10 months ago
- Recent LLM-based CV and related works. Welcome to comment/contribute!☆869Updated 4 months ago
- A curated list of papers and resources related to Described Object Detection, Open-Vocabulary/Open-World Object Detection and Referring E…☆301Updated 2 weeks ago
- Open-vocabulary Semantic Segmentation☆351Updated 9 months ago
- ☆203Updated last year
- (TPAMI 2024) A Survey on Open Vocabulary Learning☆945Updated 4 months ago
- [CVPR 2024 🔥] Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses tha…☆903Updated last month
- A curated list of awesome self-supervised learning methods in videos☆149Updated 3 weeks ago
- Reading list for research topics in Masked Image Modeling☆336Updated 8 months ago
- A collection of literature after or concurrent with Masked Autoencoder (MAE) (Kaiming He el al.).☆843Updated last year
- [T-PAMI-2024] Transformer-Based Visual Segmentation: A Survey☆745Updated 11 months ago
- A collection of project, papers, and source code for Meta AI's Segment Anything Model (SAM) and related studies.☆357Updated 7 months ago
- [CVPR 2022] Official code for "RegionCLIP: Region-based Language-Image Pretraining"☆779Updated last year
- [Survey] Masked Modeling for Self-supervised Representation Learning on Vision and Beyond (https://arxiv.org/abs/2401.00897)☆338Updated 3 months ago
- Connecting segment-anything's output masks with the CLIP model; Awesome-Segment-Anything-Works☆196Updated 9 months ago
- CVPR 2023-2024 Papers: Dive into advanced research presented at the leading computer vision conference. Keep up to date with the latest d…☆451Updated last year
- ❄️🔥 Visual Prompt Tuning [ECCV 2022] https://arxiv.org/abs/2203.12119☆1,143Updated last year
- [ICCV 2023] Official implementation of the paper "A Simple Framework for Open-Vocabulary Segmentation and Detection"☆724Updated last year
- A Survey on multimodal learning research.☆329Updated last year
- VisionLLM Series☆1,094Updated 5 months ago