LHBuilder / SA-Segment-Anything
Vision-oriented multimodal AI
☆49Updated 10 months ago
Alternatives and similar repositories for SA-Segment-Anything:
Users that are interested in SA-Segment-Anything are comparing it to the libraries listed below
- "Towards Improving Document Understanding: An Exploration on Text-Grounding via MLLMs" 2023☆14Updated 4 months ago
- Official Pytorch Implementation of Self-emerging Token Labeling☆33Updated last year
- MM-Instruct: Generated Visual Instructions for Large Multimodal Model Alignment☆34Updated 9 months ago
- Multimodal Open-O1 (MO1) is designed to enhance the accuracy of inference models by utilizing a novel prompt-based approach. This tool wo…☆29Updated 6 months ago
- ☆19Updated last year
- Codes for ICML 2023 Learning Dynamic Query Combinations for Transformer-based Object Detection and Segmentation☆37Updated last year
- Detectron2 Toolbox and Benchmark for V3Det☆16Updated 10 months ago
- LAVIS - A One-stop Library for Language-Vision Intelligence☆47Updated 8 months ago
- Evaluate the performance of computer vision models and prompts for zero-shot models (Grounding DINO, CLIP, BLIP, DINOv2, ImageBind, model…☆35Updated last year
- Empirical Study Towards Building An Effective Multi-Modal Large Language Model☆23Updated last year
- Lion: Kindling Vision Intelligence within Large Language Models☆52Updated last year
- ☆73Updated last year
- [AAAI2025] ChatterBox: Multi-round Multimodal Referring and Grounding, Multimodal, Multi-round dialogues☆53Updated 3 months ago
- ☆33Updated last year
- Pytorch implementation of HyperLLaVA: Dynamic Visual and Language Expert Tuning for Multimodal Large Language Models☆28Updated last year
- [NeurIPS2022] This is the official implementation of the paper "Expediting Large-Scale Vision Transformer for Dense Prediction without Fi…☆83Updated last year
- Detectron2 is a platform for object detection, segmentation and other visual recognition tasks.☆18Updated 2 years ago
- [NeurIPS-24] This is the official implementation of the paper "DeepStack: Deeply Stacking Visual Tokens is Surprisingly Simple and Effect…☆35Updated 9 months ago
- MLLM-DataEngine: An Iterative Refinement Approach for MLLM☆46Updated 10 months ago
- INF-LLaVA: Dual-perspective Perception for High-Resolution Multimodal Large Language Model☆42Updated 8 months ago
- OpenVLThinker: An Early Exploration to Vision-Language Reasoning via Iterative Self-Improvement☆69Updated 2 weeks ago
- Codebase for the Recognize Anything Model (RAM)☆77Updated last year
- [ICLR2025] Draw-and-Understand: Leveraging Visual Prompts to Enable MLLMs to Comprehend What You Want☆69Updated 2 months ago
- A Framework for Decoupling and Assessing the Capabilities of VLMs☆41Updated 9 months ago
- ☆68Updated 9 months ago
- Precision Search through Multi-Style Inputs☆69Updated 8 months ago
- [ECCV 2024] SegVG: Transferring Object Bounding Box to Segmentation for Visual Grounding☆56Updated 5 months ago
- ECCV2024_Parrot Captions Teach CLIP to Spot Text☆65Updated 7 months ago
- [EMNLP 2024] Official code for "Beyond Embeddings: The Promise of Visual Table in Multi-Modal Models"☆17Updated 5 months ago
- This is the official repo for ByteVideoLLM/Dynamic-VLM☆20Updated 3 months ago