autodistill / autodistill-metaclipLinks
MetaCLIP module for use with Autodistill.
☆22Updated last year
Alternatives and similar repositories for autodistill-metaclip
Users that are interested in autodistill-metaclip are comparing it to the libraries listed below
Sorting:
- Evaluate the performance of computer vision models and prompts for zero-shot models (Grounding DINO, CLIP, BLIP, DINOv2, ImageBind, model…☆37Updated 2 years ago
- Unofficial implementation and experiments related to Set-of-Mark (SoM) 👁️☆87Updated 2 years ago
- ☆15Updated last year
- ☆69Updated last year
- Testing and evaluating the capabilities of Vision-Language models (PaliGemma) in performing computer vision tasks such as object detectio…☆84Updated last year
- EdgeSAM model for use with Autodistill.☆29Updated last year
- An plug in and play pipeline that utilizes segment anything to segment datasets with rich detail for downstream fine-tuning on vision mod…☆19Updated last year
- YOLOExplorer : Iterate on your YOLO / CV datasets using SQL, Vector semantic search, and more within seconds☆136Updated 2 weeks ago
- Use Florence 2 to auto-label data for use in training fine-tuned object detection models.☆67Updated last year
- Fine-tuning OpenAI CLIP Model for Image Search on medical images☆77Updated 3 years ago
- ☆53Updated last year
- Finetune any model on HF in less than 30 seconds☆55Updated last month
- Maybe the new state of the art vision model? we'll see 🤷♂️☆165Updated last year
- ☆63Updated last year
- Visual RAG using less than 300 lines of code.☆29Updated last year
- Cerule - A Tiny Mighty Vision Model☆67Updated 2 weeks ago
- ☆50Updated 2 years ago
- Official Pytorch Implementation of Self-emerging Token Labeling☆35Updated last year
- A public implementation of the ReLoRA pretraining method, built on Lightning-AI's Pytorch Lightning suite.☆35Updated last year
- Simple Implementation of TinyGPTV in super simple Zeta lego blocks☆15Updated last year
- GroundedSAM Base Model plugin for Autodistill☆52Updated last year
- Implementation of VisionLLaMA from the paper: "VisionLLaMA: A Unified LLaMA Interface for Vision Tasks" in PyTorch and Zeta☆16Updated last year
- My personal implementation of the model from "Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities", they haven't rel…☆11Updated last year
- EfficientSAM + YOLO World base model for use with Autodistill.☆10Updated last year
- Pixel Parsing. A reproduction of OCR-free end-to-end document understanding models with open data☆22Updated last year
- ☆49Updated 9 months ago
- XmodelLM☆38Updated last year
- Official implementation of "Gemini in Reasoning: Unveiling Commonsense in Multimodal Large Language Models"☆37Updated last year
- ☆35Updated 2 years ago
- ☆59Updated last year