SkalskiP / top-cvpr-2024-papersLinks
This repository is a curated collection of the most exciting and influential CVPR 2024 papers. π₯ [Paper + Code + Demo]
β741Updated 4 months ago
Alternatives and similar repositories for top-cvpr-2024-papers
Users that are interested in top-cvpr-2024-papers are comparing it to the libraries listed below
Sorting:
- About This repository is a curated collection of the most exciting and influential CVPR 2025 papers. π₯ [Paper + Code + Demo]β795Updated 4 months ago
- Official code for "FeatUp: A Model-Agnostic Frameworkfor Features at Any Resolution" ICLR 2024β1,592Updated last year
- Official repository for "AM-RADIO: Reduce All Domains Into One"β1,360Updated last week
- RobustSAM: Segment Anything Robustly on Degraded Images (CVPR 2024 Highlight)β361Updated last year
- ποΈ + π¬ + π§ = π€ Curated list of top foundation and multimodal models! [Paper + Code + Examples + Tutorials]β636Updated last year
- Grounding DINO 1.5: IDEA Research's Most Capable Open-World Object Detection Model Seriesβ1,044Updated 9 months ago
- State-of-the-art Image & Video CLIP, Multimodal Large Language Models, and More!β1,680Updated last month
- Hiera: A fast, powerful, and simple hierarchical vision transformer.β1,026Updated last year
- π€© An AWESOME Curated List of Papers, Workshops, Datasets, and Challenges from CVPR 2024β143Updated last year
- This repository is a curated collection of the most exciting and influential CVPR 2023 papers. π₯ [Paper + Code]β653Updated 4 months ago
- Efficient Track Anythingβ656Updated 9 months ago
- This repo is the homebase of a community driven course on Computer Vision with Neural Networks. Feel free to join us on the Hugging Face β¦β723Updated last month
- code for CVPR2024 paper: DiffMOT: A Real-time Diffusion-based Multiple Object Tracker with Non-linear Predictionβ433Updated last year
- 4M: Massively Multimodal Masked Modelingβ1,764Updated 4 months ago
- A curated list of foundation models for vision and language tasksβ1,102Updated 4 months ago
- β528Updated 11 months ago
- [NeurIPS 2024] Code release for "Segment Anything without Supervision"β485Updated last month
- Tracking Any Point (TAP)β1,691Updated last week
- β62Updated 2 years ago
- All-in-one training for vision models (YOLO, ViTs, RT-DETR, DINOv3): pretraining, fine-tuning, distillation.β991Updated this week
- Recipes for shrinking, optimizing, customizing cutting edge vision models. πβ1,642Updated last month
- [CVPR 2024 π₯] Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses thaβ¦β922Updated 2 months ago
- [CVPR 2025 Highlight] Official repository for the paper: "SAMWISE: Infusing Wisdom in SAM2 for Text-Driven Video Segmentation"β331Updated 3 weeks ago
- ICLR2024 Spotlight: curation/training code, metadata, distribution and pre-trained models for MetaCLIP; CVPR 2024: MoDE: CLIP Data Expertβ¦β1,694Updated 3 weeks ago
- This repository provides the code and model checkpoints for AIMv1 and AIMv2 research projects.β1,377Updated 2 months ago
- [ICCV 2023] Tracking Anything with Decoupled Video Segmentationβ1,438Updated 5 months ago
- A collection of papers on the topic of ``Computer Vision in the Wild (CVinW)''β1,339Updated last year
- [CVPRW'24] SoccerNet Game State Reconstruction: End-to-End Athlete Tracking and Identification on a Minimap (CVPR24 - CVSports workshop)β339Updated 5 months ago
- [CVPR 2025 Highlight] Official code and models for Encoder-only Mask Transformer (EoMT).β439Updated this week
- SAM with text promptβ2,421Updated last month