TencentARC / ARC-Hunyuan-Video-7BLinks
Structured Video Comprehension of Real-World Shorts
☆230Updated 4 months ago
Alternatives and similar repositories for ARC-Hunyuan-Video-7B
Users that are interested in ARC-Hunyuan-Video-7B are comparing it to the libraries listed below
Sorting:
- Ego-R1: Chain-of-Tool-Thought for Ultra-Long Egocentric Video Reasoning☆137Updated 5 months ago
- ☆175Updated 7 months ago
- [ICCV2025]Code Release of Harmonizing Visual Representations for Unified Multimodal Understanding and Generation☆185Updated 8 months ago
- Empowering Unified MLLM with Multi-granular Visual Generation☆129Updated last year
- [ICLR 2025] AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark☆138Updated 8 months ago
- Video-Holmes: Can MLLM Think Like Holmes for Complex Video Reasoning?☆86Updated 6 months ago
- ICML2025☆63Updated 5 months ago
- Official repository for the UAE paper, unified-GRPO, and unified-Bench☆156Updated 4 months ago
- ☆141Updated 3 months ago
- [NeurIPS 2024 D&B Track] Official Repo for "LVD-2M: A Long-take Video Dataset with Temporally Dense Captions"☆76Updated last year
- [ICLR 2026] This is an early exploration to introduce Interleaving Reasoning to Text-to-image Generation field and achieve the SoTA bench…☆86Updated last week
- GoT-R1: Unleashing Reasoning Capability of MLLM for Visual Generation with Reinforcement Learning☆101Updated last week
- WISE: A World Knowledge-Informed Semantic Evaluation for Text-to-Image Generation☆182Updated 2 months ago
- UniVG-R1: Reasoning Guided Universal Visual Grounding with Reinforcement Learning☆156Updated 8 months ago
- ☆96Updated 7 months ago
- Code release for Ming-UniVision: Joint Image Understanding and Geneation with a Continuous Unified Tokenizer☆136Updated 3 months ago
- [NIPS2025] VideoChat-R1 & R1.5: Enhancing Spatio-Temporal Perception and Reasoning via Reinforcement Fine-Tuning☆255Updated 3 months ago
- The official code of "Thinking With Videos: Multimodal Tool-Augmented Reinforcement Learning for Long Video Reasoning"☆79Updated 3 months ago
- [NeurlPS 2024] One Token to Seg Them All: Language Instructed Reasoning Segmentation in Videos☆145Updated last year
- TokLIP: Marry Visual Tokens to CLIP for Multimodal Comprehension and Generation☆236Updated 5 months ago
- https://huggingface.co/datasets/multimodal-reasoning-lab/Zebra-CoT☆117Updated this week
- [arXiv: 2502.05178] QLIP: Text-Aligned Visual Tokenization Unifies Auto-Regressive Multimodal Understanding and Generation☆95Updated 11 months ago
- [NIPS 25'] Evaluation code of paper "KRIS-Bench: Benchmarking Next-Level Intelligent Image Editing Models"☆39Updated 3 months ago
- Code and dataset link for "DenseWorld-1M: Towards Detailed Dense Grounded Caption in the Real World"☆122Updated 4 months ago
- [ICLR'25] Reconstructive Visual Instruction Tuning☆135Updated 9 months ago
- TimeLens: Rethinking Video Temporal Grounding with Multimodal LLMs☆97Updated last week
- ☆123Updated 5 months ago
- [CVPR 2025] OVO-Bench: How Far is Your Video-LLMs from Real-World Online Video Understanding?☆118Updated 6 months ago
- LongVT: Incentivizing "Thinking with Long Videos" via Native Tool Calling☆186Updated last week
- [NeurIPS 2025 Spotlight] A Unified Tokenizer for Visual Generation and Understanding☆503Updated 2 months ago