sanbuphy / computer-vision-referenceLinks
Collected the world's best computer vision labs and lecture materials.
☆14Updated 11 months ago
Alternatives and similar repositories for computer-vision-reference
Users that are interested in computer-vision-reference are comparing it to the libraries listed below
Sorting:
- ☆137Updated last week
- Rethinking RL Scaling for Vision Language Models: A Transparent, From-Scratch Framework and Comprehensive Evaluation Scheme☆147Updated 9 months ago
- ☆168Updated 2 months ago
- A Collection of Papers on Diffusion Language Models☆154Updated 4 months ago
- G1: Bootstrapping Perception and Reasoning Abilities of Vision-Language Model via Reinforcement Learning☆95Updated 8 months ago
- The official repository for the paper "ThinkMorph: Emergent Properties in Multimodal Interleaved Chain-of-Thought Reasoning"☆140Updated 3 weeks ago
- [ArXiv] V2PE: Improving Multimodal Long-Context Capability of Vision-Language Models with Variable Visual Position Encoding☆59Updated last year
- Doodling our way to AGI ✏️ 🖼️ 🧠☆120Updated 8 months ago
- The official code of "VL-Rethinker: Incentivizing Self-Reflection of Vision-Language Models with Reinforcement Learning" [NeurIPS25]☆179Updated 7 months ago
- A collection of papers on discrete diffusion models☆168Updated 7 months ago
- A Massive Multi-Discipline Lecture Understanding Benchmark☆32Updated 3 months ago
- VideoNSA: Native Sparse Attention Scales Video Understanding☆79Updated 2 months ago
- Evaluating Knowledge Acquisition from Multi-Discipline Professional Videos☆64Updated 4 months ago
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆109Updated 8 months ago
- ☆34Updated 5 months ago
- This is the official repository for the paper "MathCanvas: Intrinsic Visual Chain-of-Thought for Multimodal Mathematical Reasoning"☆57Updated last month
- ✨✨ [ICLR 2026] MME-Unify: A Comprehensive Benchmark for Unified Multimodal Understanding and Generation Models☆42Updated 9 months ago
- Minimalist RL for Diffusion LLMs with SOTA reasoning performance (89.1% GSM8K). Official implementation of "The Flexibility Trap".☆75Updated last week
- GroundCUA☆65Updated last month
- Data and Code for CVPR 2025 paper "MMVU: Measuring Expert-Level Multi-Discipline Video Understanding"☆77Updated 11 months ago
- ☆204Updated last month
- paper list, tutorial, and nano code snippet for Diffusion Large Language Models.☆152Updated 2 weeks ago
- [CVPR2025] BOLT: Boost Large Vision-Language Model Without Training for Long-form Video Understanding☆37Updated 10 months ago
- Recent Advances on MLLM's Reasoning Ability☆26Updated 9 months ago
- ☆110Updated last year
- ☆20Updated 8 months ago
- https://huggingface.co/datasets/multimodal-reasoning-lab/Zebra-CoT☆117Updated 3 months ago
- Thinking with Videos from Open-Source Priors. We reproduce chain-of-frames visual reasoning by fine-tuning open-source video models. Give…☆206Updated 3 months ago
- Locality-aware Parallel Decoding for Efficient Autoregressive Image Generation☆81Updated 6 months ago
- [ICLR'26] Traceable Evidence Enhanced Visual Grounded Reasoning: Evaluation and Methodology☆72Updated this week