roboflow / cog-vlm-clientLinks
Simple CogVLM client script
β14Updated last year
Alternatives and similar repositories for cog-vlm-client
Users that are interested in cog-vlm-client are comparing it to the libraries listed below
Sorting:
- Unofficial implementation and experiments related to Set-of-Mark (SoM) ποΈβ86Updated last year
- EdgeSAM model for use with Autodistill.β27Updated last year
- β29Updated last year
- Evaluate the performance of computer vision models and prompts for zero-shot models (Grounding DINO, CLIP, BLIP, DINOv2, ImageBind, modelβ¦β36Updated last year
- β14Updated last year
- BUD-E (Buddy) is an open-source voice assistant framework that facilitates seamless interaction with AI models and APIs, enabling the creβ¦β20Updated 9 months ago
- β20Updated last year
- GPT-4V(ision) module for use with Autodistill.β26Updated 11 months ago
- A collection of notebooks for the Hugging Face blog series (https://huggingface.co/blog).β45Updated 11 months ago
- BH hackathonβ14Updated last year
- Testing and evaluating the capabilities of Vision-Language models (PaliGemma) in performing computer vision tasks such as object detectioβ¦β81Updated last year
- Build Agentic workflows with function calling using open LLMsβ28Updated last week
- Visual RAG using less than 300 lines of code.β28Updated last year
- Use Grounding DINO, Segment Anything, and GPT-4V to label images with segmentation masks for use in training smaller, fine-tuned models.β66Updated last year
- Use Florence 2 to auto-label data for use in training fine-tuned object detection models.β64Updated 11 months ago
- Demo python script app to interact with llama.cpp server using whisper API, microphone and webcam devices.β46Updated last year
- Cerule - A Tiny Mighty Vision Modelβ66Updated 10 months ago
- AgentParse is a high-performance parsing library designed to map various structured data formats (such as Pydantic models, JSON, YAML, anβ¦β13Updated 2 weeks ago
- β16Updated last year
- My personal implementation of the model from "Qwen-VL: A Frontier Large Vision-Language Model with Versatile Abilities", they haven't relβ¦β13Updated last year
- Take your LLM to the optometrist.β32Updated last week
- Finetune any model on HF in less than 30 secondsβ57Updated 3 months ago
- β68Updated last year
- Use Grounding DINO, Segment Anything, and CLIP to label objects in images.β31Updated last year
- Passively collect images for computer vision datasets on the edge.β34Updated last year
- Command-line script for inferencing from models such as WizardCoderβ26Updated last year
- β50Updated last year
- Implementation of VisionLLaMA from the paper: "VisionLLaMA: A Unified LLaMA Interface for Vision Tasks" in PyTorch and Zetaβ16Updated 8 months ago
- Pixel Parsing. A reproduction of OCR-free end-to-end document understanding models with open dataβ21Updated 11 months ago
- β14Updated last year