SkalskiP / top-cvpr-2024-papers
This repository is a curated collection of the most exciting and influential CVPR 2024 papers. π₯ [Paper + Code + Demo]
β711Updated 9 months ago
Alternatives and similar repositories for top-cvpr-2024-papers:
Users that are interested in top-cvpr-2024-papers are comparing it to the libraries listed below
- Official repository for "AM-RADIO: Reduce All Domains Into One"β1,105Updated this week
- Grounding DINO 1.5: IDEA Research's Most Capable Open-World Object Detection Model Seriesβ930Updated 2 months ago
- RobustSAM: Segment Anything Robustly on Degraded Images (CVPR 2024 Highlight)β347Updated 7 months ago
- Official Implementation of CVPR24 highlight paper: Matching Anything by Segmenting Anythingβ1,262Updated 5 months ago
- Official code for "FeatUp: A Model-Agnostic Frameworkfor Features at Any Resolution" ICLR 2024β1,487Updated 9 months ago
- Hiera: A fast, powerful, and simple hierarchical vision transformer.β971Updated last year
- code for CVPR2024 paper: DiffMOT: A Real-time Diffusion-based Multiple Object Tracker with Non-linear Predictionβ407Updated 10 months ago
- ποΈ + π¬ + π§ = π€ Curated list of top foundation and multimodal models! [Paper + Code + Examples + Tutorials]β611Updated last year
- Combining Segment Anything (SAM) with Grounded DINO for zero-shot object detection and CLIPSeg for zero-shot segmentationβ404Updated 11 months ago
- This repository is a curated collection of the most exciting and influential CVPR 2023 papers. π₯ [Paper + Code]β646Updated 9 months ago
- DINO-X: The World's Top-Performing Vision Model for Open-World Object Detection and Understandingβ998Updated 3 weeks ago
- [NeurIPS 2024] Code release for "Segment Anything without Supervision"β460Updated 6 months ago
- π€© An AWESOME Curated List of Papers, Workshops, Datasets, and Challenges from CVPR 2024β142Updated 10 months ago
- Grounded SAM 2: Ground and Track Anything in Videos with Grounding DINO, Florence-2 and SAM 2β1,990Updated 3 weeks ago
- Recipes for shrinking, optimizing, customizing cutting edge vision models. πβ1,407Updated 3 weeks ago
- Official code of "EVF-SAM: Early Vision-Language Fusion for Text-Prompted Segment Anything Model"β390Updated last month
- This repo is the homebase of a community driven course on Computer Vision with Neural Networks. Feel free to join us on the Hugging Face β¦β619Updated this week
- β507Updated 5 months ago
- 4M: Massively Multimodal Masked Modelingβ1,713Updated last month
- Efficient Track Anythingβ525Updated 3 months ago
- [CVPR 2024 π₯] Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses thaβ¦β863Updated 4 months ago
- This series will take you on a journey from the fundamentals of NLP and Computer Vision to the cutting edge of Vision-Language Models.β1,061Updated 2 months ago
- Famous Vision Language Models and Their Architecturesβ778Updated last month
- [ICCV 2023] Tracking Anything with Decoupled Video Segmentationβ1,357Updated 8 months ago
- [CVPR 2023] Official implementation of the paper "Mask DINO: Towards A Unified Transformer-based Framework for Object Detection and Segmeβ¦β1,303Updated last year
- Images to inference with no labeling (use foundation models to train supervised models).β2,210Updated 3 weeks ago
- This repository is for the first comprehensive survey on Meta AI's Segment Anything Model (SAM).β938Updated this week
- [ICLR 2024] Official PyTorch implementation of FasterViT: Fast Vision Transformers with Hierarchical Attentionβ843Updated 3 weeks ago
- SAM with text promptβ2,108Updated 2 months ago
- Official Pytorch Implementation for βDINO-Tracker: Taming DINO for Self-Supervised Point Tracking in a Single Videoβ (ECCV 2024)β477Updated 4 months ago