SkalskiP / top-cvpr-2024-papers
This repository is a curated collection of the most exciting and influential CVPR 2024 papers. π₯ [Paper + Code + Demo]
β685Updated 6 months ago
Alternatives and similar repositories for top-cvpr-2024-papers:
Users that are interested in top-cvpr-2024-papers are comparing it to the libraries listed below
- RobustSAM: Segment Anything Robustly on Degraded Images (CVPR 2024 Highlight)β331Updated 4 months ago
- Official repository for "AM-RADIO: Reduce All Domains Into One"β892Updated this week
- ποΈ + π¬ + π§ = π€ Curated list of top foundation and multimodal models! [Paper + Code + Examples + Tutorials]β596Updated 10 months ago
- Official Implementation of CVPR24 highligt paper: Matching Anything by Segmenting Anythingβ1,152Updated 2 months ago
- π€© An AWESOME Curated List of Papers, Workshops, Datasets, and Challenges from CVPR 2024β141Updated 7 months ago
- This repository is a curated collection of the most exciting and influential CVPR 2023 papers. π₯ [Paper + Code]β645Updated 6 months ago
- Official code for "FeatUp: A Model-Agnostic Frameworkfor Features at Any Resolution" ICLR 2024β1,431Updated 6 months ago
- Grounding DINO 1.5: IDEA Research's Most Capable Open-World Object Detection Model Seriesβ853Updated 5 months ago
- β481Updated 2 months ago
- [NeurIPS 2024] Code release for "Segment Anything without Supervision"β441Updated 3 months ago
- This series will take you on a journey from the fundamentals of NLP and Computer Vision to the cutting edge of Vision-Language Models.β720Updated last week
- Hiera: A fast, powerful, and simple hierarchical vision transformer.β940Updated 10 months ago
- This repo is the homebase of a community driven course on Computer Vision with Neural Networks. Feel free to join us on the Hugging Face β¦β536Updated this week
- A curated list of foundation models for vision and language tasksβ905Updated this week
- Efficient Track Anythingβ441Updated last week
- streamline the fine-tuning process for multimodal models: PaliGemma, Florence-2, and Qwen2-VLβ1,427Updated this week
- code for CVPR2024 paper: DiffMOT: A Real-time Diffusion-based Multiple Object Tracker with Non-linear Predictionβ400Updated 7 months ago
- CVPR 2023-2024 Papers: Dive into advanced research presented at the leading computer vision conference. Keep up to date with the latest dβ¦β438Updated 6 months ago
- A collection of papers on the topic of ``Computer Vision in the Wild (CVinW)''β1,237Updated 10 months ago
- 4M: Massively Multimodal Masked Modelingβ1,666Updated 3 months ago
- Grounded SAM 2: Ground and Track Anything in Videos with Grounding DINO, Florence-2 and SAM 2β1,518Updated 3 weeks ago
- A distilled Segment Anything (SAM) model capable of running real-time with NVIDIA TensorRTβ696Updated last year
- Open source AI/ML capabilities for the FiftyOne ecosystemβ134Updated this week
- This repository contains the official implementation of the research paper, "MobileCLIP: Fast Image-Text Models through Multi-Modal Reinfβ¦β792Updated last month
- Recipes for shrinking, optimizing, customizing cutting edge vision models. πβ1,093Updated 3 weeks ago
- [ICCV 2023] Tracking Anything with Decoupled Video Segmentationβ1,302Updated 5 months ago
- This repository provides the code and model checkpoints for AIMv1 and AIMv2 research projects.β1,143Updated last month
- Famous Vision Language Models and Their Architecturesβ565Updated 4 months ago
- Official codebase used to develop Vision Transformer, SigLIP, MLP-Mixer, LiT and more.β2,516Updated 3 weeks ago
- [CVPR 2024 π₯] Grounding Large Multimodal Model (GLaMM), the first-of-its-kind model capable of generating natural language responses thaβ¦β814Updated last month