Aasthaengg / GLIP-BLIP-Vision-Langauge-Obj-Det-VQA
☆32Updated 2 years ago
Alternatives and similar repositories for GLIP-BLIP-Vision-Langauge-Obj-Det-VQA:
Users that are interested in GLIP-BLIP-Vision-Langauge-Obj-Det-VQA are comparing it to the libraries listed below
- Official repository for the General Robust Image Task (GRIT) Benchmark☆51Updated last year
- Evaluate the performance of computer vision models and prompts for zero-shot models (Grounding DINO, CLIP, BLIP, DINOv2, ImageBind, model…☆35Updated last year
- A simple wrapper library for binding timm models as detectron2 backbones☆39Updated last year
- Official PyTorch implementation of RIO☆18Updated 3 years ago
- [FGVC9-CVPR 2022] The second place solution for 2nd eBay eProduct Visual Search Challenge.☆26Updated 2 years ago
- ☆88Updated last year
- Official Pytorch Implementation of Self-emerging Token Labeling☆32Updated 11 months ago
- Official Training and Inference Code of Amodal Expander, Proposed in Tracking Any Object Amodally☆15Updated 7 months ago
- [AAAI2025] ChatterBox: Multi-round Multimodal Referring and Grounding, Multimodal, Multi-round dialogues☆53Updated 2 months ago
- PyTorch implementation of "UNIT: Unifying Image and Text Recognition in One Vision Encoder", NeurlPS 2024.☆27Updated 5 months ago
- Code for Recall@k Surrogate Loss with Large Batches and Similarity Mixup, CVPR 2022.☆60Updated 4 months ago
- A pytorch Implementation of Open Vocabulary Object Detection with Pseudo Bounding-Box Labels☆59Updated last year
- ☆42Updated last month
- 4th place solution for the Google Universal Image Embedding Kaggle Challenge. Instance-Level Recognition workshop at ECCV 2022☆42Updated last year
- Official repo of Griffon series including v1(ECCV 2024), v2, and G☆132Updated last month
- A task-agnostic vision-language architecture as a step towards General Purpose Vision☆92Updated 3 years ago
- ☆64Updated last year
- Our public repo ranked 1st 🏆🏆 at MMSports2023 challenge on segmentation task☆16Updated last year
- Code for experiments for "ConvNet vs Transformer, Supervised vs CLIP: Beyond ImageNet Accuracy"☆101Updated 5 months ago
- Vision-oriented multimodal AI☆49Updated 8 months ago
- Code for AAAI 2023 Paper : “Alignment-Enriched Tuning for Patch-Level Pre-trained Document Image Models”☆17Updated 2 years ago
- Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training☆135Updated 2 years ago
- A huge dataset for Document Visual Question Answering☆15Updated 7 months ago
- Use CLIP to represent video for Retrieval Task☆69Updated 4 years ago
- ZoomEye: Enhancing Multimodal LLMs with Human-Like Zooming Capabilities through Tree-Based Image Exploration☆24Updated 2 months ago
- Simple Implementation of Pix2Seq model for object detection in PyTorch☆122Updated last year
- Object Recognition as Next Token Prediction (CVPR 2024 Highlight)☆174Updated 2 months ago
- Code and Models for "GeneCIS A Benchmark for General Conditional Image Similarity"☆56Updated last year
- [ICME 2022] code for the paper, SimVit: Exploring a simple vision transformer with sliding windows.☆67Updated 2 years ago
- [BMVC22] Official Implementation of ViCHA: "Efficient Vision-Language Pretraining with Visual Concepts and Hierarchical Alignment"☆54Updated 2 years ago