yh-hust / VisuRiddlesLinks
VisuRiddles: Fine-grained Perception is a important thing for Multimodal Large Models in Riddles Solving
☆17Updated last month
Alternatives and similar repositories for VisuRiddles
Users that are interested in VisuRiddles are comparing it to the libraries listed below
Sorting:
- R1-Vision: Let's first take a look at the image☆48Updated 9 months ago
- The official repo for “TextCoT: Zoom In for Enhanced Multimodal Text-Rich Image Understanding”.☆44Updated last year
- Official repo of Griffon series including v1(ECCV 2024), v2(ICCV 2025), G, and R, and also the RL tool Vision-R1.☆246Updated 4 months ago
- AAAI 2024: Visual Instruction Generation and Correction☆94Updated last year
- MM-Eureka V0 also called R1-Multimodal-Journey, Latest version is in MM-Eureka☆321Updated 5 months ago
- The proposed simulated dataset consisting of 9,536 charts and associated data annotations in CSV format.☆26Updated last year
- ☆14Updated 6 months ago
- SVIT: Scaling up Visual Instruction Tuning☆165Updated last year
- Pink: Unveiling the Power of Referential Comprehension for Multi-modal LLMs☆95Updated 10 months ago
- Latest open-source "Thinking with images" (O3/O4-mini) papers, covering training-free, SFT-based, and RL-enhanced methods for "fine-grain…☆103Updated 3 months ago
- [ICLR2025] Draw-and-Understand: Leveraging Visual Prompts to Enable MLLMs to Comprehend What You Want☆92Updated last week
- 【ArXiv】PDF-Wukong: A Large Multimodal Model for Efficient Long PDF Reading with End-to-End Sparse Sampling☆128Updated 6 months ago
- [COLM 2025] Open-Qwen2VL: Compute-Efficient Pre-Training of Fully-Open Multimodal LLMs on Academic Resources☆288Updated 3 months ago
- Harnessing 1.4M GPT4V-synthesized Data for A Lite Vision-Language Model☆277Updated last year
- [CVPR 2025] LamRA: Large Multimodal Model as Your Advanced Retrieval Assistant☆171Updated 5 months ago
- [CVPR 2024] LION: Empowering Multimodal Large Language Model with Dual-Level Visual Knowledge☆154Updated 3 months ago
- [ICLR 2025] LLaVA-MoD: Making LLaVA Tiny via MoE-Knowledge Distillation☆211Updated 8 months ago
- ☆22Updated 11 months ago
- ☆356Updated last year
- [ICLR 2025 Spotlight] OmniCorpus: A Unified Multimodal Corpus of 10 Billion-Level Images Interleaved with Text☆408Updated 7 months ago
- ☆142Updated last year
- Official Repo of "MMBench: Is Your Multi-modal Model an All-around Player?"☆275Updated 6 months ago
- [NeurIPS 2024] This repo contains evaluation code for the paper "Are We on the Right Way for Evaluating Large Vision-Language Models"☆200Updated last year
- Turning a CLIP Model into a Scene Text Detector (CVPR2023) | Turning a CLIP Model into a Scene Text Spotter (TPAMI)☆197Updated last year
- ☆46Updated 10 months ago
- A collection of visual instruction tuning datasets.☆76Updated last year
- The official code for NeurIPS 2024 paper: Harmonizing Visual Text Comprehension and Generation☆129Updated last year
- A bug-free and improved implementation of LLaVA-UHD, based on the code from the official repo☆34Updated last year
- [ECCV 2024] Official PyTorch implementation of DreamLIP: Language-Image Pre-training with Long Captions☆136Updated 7 months ago
- VoCoT: Unleashing Visually Grounded Multi-Step Reasoning in Large Multi-Modal Models☆77Updated last year