zhengrongz / AoTD
Official PyTorch code of "Enhancing Video-LLM Reasoning via Agent-of-Thoughts Distillation".
☆22Updated last month
Alternatives and similar repositories for AoTD:
Users that are interested in AoTD are comparing it to the libraries listed below
- [AAAI 2025] Grounded Multi-Hop VideoQA in Long-Form Egocentric Videos☆23Updated 6 months ago
- Official PyTorch code of GroundVQA (CVPR'24)☆58Updated 6 months ago
- Envolving Temporal Reasoning Capability into LMMs via Temporal Consistent Reward☆20Updated last week
- ☆23Updated 6 months ago
- Official implementation of HawkEye: Training Video-Text LLMs for Grounding Text in Videos☆40Updated 11 months ago
- ☆28Updated 6 months ago
- Latest Advances on (RL based) Multimodal Reasoning and Generation in Multimodal Large Language Models☆17Updated this week
- R1-like Video-LLM for Temporal Grounding☆62Updated last week
- ☆86Updated 3 months ago
- ☆16Updated 4 months ago
- A Versatile Video-LLM for Long and Short Video Understanding with Superior Temporal Localization Ability☆90Updated 4 months ago
- [CVPR2025] Number it: Temporal Grounding Videos like Flipping Manga☆67Updated this week
- [CVPR 2025] Adaptive Keyframe Sampling for Long Video Understanding☆42Updated last week
- [CVPR'25] VideoEspresso: A Large-Scale Chain-of-Thought Dataset for Fine-Grained Video Reasoning via Core Frame Selection☆63Updated last week
- Can I Trust Your Answer? Visually Grounded Video Question Answering (CVPR'24, Highlight)☆66Updated 9 months ago
- [CVPR 2024] Context-Guided Spatio-Temporal Video Grounding☆51Updated 9 months ago
- The official repository for paper "PruneVid: Visual Token Pruning for Efficient Video Large Language Models".☆35Updated last month
- Code for paper "VideoTree: Adaptive Tree-based Video Representation for LLM Reasoning on Long Videos"☆102Updated last month
- [Open LLaVA-Video-R1]✨First Adaptation of R1 to LLaVA-Video☆24Updated last week
- LongVALE: Vision-Audio-Language-Event Benchmark Towards Time-Aware Omni-Modal Perception of Long Videos. (CVPR 2025))☆18Updated last week
- (CVPR 2025) PyramidDrop: Accelerating Your Large Vision-Language Models via Pyramid Visual Redundancy Reduction☆84Updated 3 weeks ago
- PyTorch code for "Contrastive Region Guidance: Improving Grounding in Vision-Language Models without Training"☆33Updated last year
- [ICLR 2025] TimeSuite: Improving MLLMs for Long Video Understanding via Grounded Tuning☆28Updated 2 weeks ago
- [ICLR 2025] TRACE: Temporal Grounding Video LLM via Casual Event Modeling☆76Updated 2 months ago
- This is the official implementation of ReVisionLLM: Recursive Vision-Language Model for Temporal Grounding in Hour-Long Videos☆11Updated this week
- p-MoD: Building Mixture-of-Depths MLLMs via Progressive Ratio Decay☆33Updated 2 months ago
- [CVPR 2024] Retrieval-Augmented Image Captioning with External Visual-Name Memory for Open-World Comprehension☆48Updated 11 months ago
- VideoNIAH: A Flexible Synthetic Method for Benchmarking Video MLLMs☆46Updated 3 weeks ago
- HallE-Control: Controlling Object Hallucination in LMMs☆30Updated 11 months ago
- [EMNLP 2024 Findings] The official PyTorch implementation of EchoSight: Advancing Visual-Language Models with Wiki Knowledge.☆55Updated 2 weeks ago