roboflow / cvevalsLinks
Evaluate the performance of computer vision models and prompts for zero-shot models (Grounding DINO, CLIP, BLIP, DINOv2, ImageBind, models hosted on Roboflow)
☆36Updated last year
Alternatives and similar repositories for cvevals
Users that are interested in cvevals are comparing it to the libraries listed below
Sorting:
- EdgeSAM model for use with Autodistill.☆29Updated last year
- Vision-oriented multimodal AI☆49Updated last year
- ☆69Updated last year
- Official Pytorch Implementation of Self-emerging Token Labeling☆35Updated last year
- Implementation of VisionLLaMA from the paper: "VisionLLaMA: A Unified LLaMA Interface for Vision Tasks" in PyTorch and Zeta☆16Updated 10 months ago
- Use Florence 2 to auto-label data for use in training fine-tuned object detection models.☆67Updated last year
- Unofficial implementation and experiments related to Set-of-Mark (SoM) 👁️☆88Updated last year
- YOLOExplorer : Iterate on your YOLO / CV datasets using SQL, Vector semantic search, and more within seconds☆135Updated 2 weeks ago
- Timm model explorer☆41Updated last year
- ☆59Updated last year
- [NeurIPS 2023] HASSOD: Hierarchical Adaptive Self-Supervised Object Detection☆58Updated last year
- EfficientSAM + YOLO World base model for use with Autodistill.☆10Updated last year
- Use Segment Anything 2, grounded with Florence-2, to auto-label data for use in training vision models.☆128Updated last year
- GroundedSAM Base Model plugin for Autodistill☆52Updated last year
- Pixel Parsing. A reproduction of OCR-free end-to-end document understanding models with open data☆21Updated last year
- Use Grounding DINO, Segment Anything, and GPT-4V to label images with segmentation masks for use in training smaller, fine-tuned models.☆66Updated last year
- A component that allows you to annotate an image with points and boxes.☆21Updated last year
- Code for experiments for "ConvNet vs Transformer, Supervised vs CLIP: Beyond ImageNet Accuracy"☆101Updated last year
- ☆15Updated last year
- Use Grounding DINO, Segment Anything, and CLIP to label objects in images.☆31Updated last year
- Implementation of the model: "(MC-ViT)" from the paper: "Memory Consolidation Enables Long-Context Video Understanding"☆23Updated last week
- SAM-CLIP module for use with Autodistill.☆15Updated last year
- Fine-tuning OpenAI CLIP Model for Image Search on medical images☆77Updated 3 years ago
- ☆79Updated 10 months ago
- ☆87Updated last year
- Tracking through Containers and Occluders in the Wild (CVPR 2023) - Official Implementation☆41Updated last year
- OLA-VLM: Elevating Visual Perception in Multimodal LLMs with Auxiliary Embedding Distillation, arXiv 2024☆61Updated 6 months ago
- ☆24Updated 11 months ago
- Pytorch implementation of HyperLLaVA: Dynamic Visual and Language Expert Tuning for Multimodal Large Language Models☆28Updated last year
- ClickDiffusion: Harnessing LLMs for Interactive Precise Image Editing☆70Updated last year