ali-vilab / CAPabilityLinks
What Is a Good Caption? A Comprehensive Visual Caption Benchmark for Evaluating Both Correctness and Thoroughness
☆26Updated 8 months ago
Alternatives and similar repositories for CAPability
Users that are interested in CAPability are comparing it to the libraries listed below
Sorting:
- [ECCV 2024] ShareGPT4V: Improving Large Multi-modal Models with Better Captions☆248Updated last year
- A Large-scale Dataset for training and evaluating model's ability on Dense Text Image Generation☆86Updated 4 months ago
- A Versatile Video-LLM for Long and Short Video Understanding with Superior Temporal Localization Ability☆105Updated last year
- ☆27Updated 9 months ago
- Code and dataset link for "DenseWorld-1M: Towards Detailed Dense Grounded Caption in the Real World"☆122Updated 4 months ago
- Structured Video Comprehension of Real-World Shorts☆230Updated 4 months ago
- [EMNLP 2025 Findings] Grounded-VideoLLM: Sharpening Fine-grained Temporal Grounding in Video Large Language Models☆139Updated 5 months ago
- [CVPR 2025] OVO-Bench: How Far is Your Video-LLMs from Real-World Online Video Understanding?☆120Updated 6 months ago
- [ICLR 2025] AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark☆138Updated 8 months ago
- LongVT: Incentivizing "Thinking with Long Videos" via Native Tool Calling☆186Updated 2 weeks ago
- TimeLens: Rethinking Video Temporal Grounding with Multimodal LLMs☆101Updated this week
- The official code of "Thinking With Videos: Multimodal Tool-Augmented Reinforcement Learning for Long Video Reasoning"☆80Updated 3 months ago
- [ICLR 2025] Diffusion Feedback Helps CLIP See Better☆299Updated last year
- Unified layout planning and image generation, ICCV2025☆40Updated 3 weeks ago
- [NeurlPS 2024] One Token to Seg Them All: Language Instructed Reasoning Segmentation in Videos☆145Updated last year
- ☆160Updated last year
- Codes for ICLR 2025 Paper: Towards Semantic Equivalence of Tokenization in Multimodal LLM☆77Updated 9 months ago
- ☆24Updated last year
- Video-Holmes: Can MLLM Think Like Holmes for Complex Video Reasoning?☆86Updated 6 months ago
- [ICLR'26] Easier Painting Than Thinking: Can Text-to-Image Models Set the Stage, but Not Direct the Play?☆48Updated last week
- ☆37Updated 7 months ago
- ☆25Updated 2 months ago
- LinVT: Empower Your Image-level Large Language Model to Understand Videos☆84Updated last year
- [NIPS 2025 DB Oral] Official Repository of paper: Envisioning Beyond the Pixels: Benchmarking Reasoning-Informed Visual Editing☆140Updated this week
- 🔥Awesome Multimodal Large Language Models Paper List☆154Updated 10 months ago
- [CVPR2025] Number it: Temporal Grounding Videos like Flipping Manga☆144Updated 3 weeks ago
- (ICLR 2026)Official repository of 'ScaleCap: Inference-Time Scalable Image Captioning via Dual-Modality Debiasing’☆58Updated last week
- [ACL 2025] PruneVid: Visual Token Pruning for Efficient Video Large Language Models☆66Updated 8 months ago
- [CVPRW 2025] UniToken is an auto-regressive generation model that combines discrete and continuous representations to process visual inpu…☆105Updated 9 months ago
- [NIPS2025] VideoChat-R1 & R1.5: Enhancing Spatio-Temporal Perception and Reasoning via Reinforcement Fine-Tuning☆256Updated 3 months ago