ruili33 / TPOLinks
☆40Updated 3 months ago
Alternatives and similar repositories for TPO
Users that are interested in TPO are comparing it to the libraries listed below
Sorting:
- [AAAI 2025] Grounded Multi-Hop VideoQA in Long-Form Egocentric Videos☆29Updated 7 months ago
- Official implement of MIA-DPO☆69Updated 11 months ago
- FreeVA: Offline MLLM as Training-Free Video Assistant☆68Updated last year
- Reinforcement Learning Tuning for VideoLLMs: Reward Design and Data Efficiency☆60Updated 6 months ago
- [NeurIPS 2025 Spotlight] Fast-Slow Thinking GRPO for Large Vision-Language Model Reasoning☆23Updated last week
- [CVPR 2025 Oral] VideoEspresso: A Large-Scale Chain-of-Thought Dataset for Fine-Grained Video Reasoning via Core Frame Selection☆131Updated 5 months ago
- (ICLR 2025 Spotlight) Official code repository for Interleaved Scene Graph.☆31Updated 4 months ago
- [ICCV 2025] Official code for "AIM: Adaptive Inference of Multi-Modal LLMs via Token Merging and Pruning"☆49Updated 2 months ago
- Code for "AVG-LLaVA: A Multimodal Large Model with Adaptive Visual Granularity"☆32Updated last year
- 🔥 [ICLR 2025] Official PyTorch Model "Visual Haystacks: A Vision-Centric Needle-In-A-Haystack Benchmark"☆26Updated 10 months ago
- Video-Holmes: Can MLLM Think Like Holmes for Complex Video Reasoning?☆84Updated 5 months ago
- TStar is a unified temporal search framework for long-form video question answering☆80Updated 3 months ago
- ☆35Updated last year
- Official repository of 'ScaleCap: Inference-Time Scalable Image Captioning via Dual-Modality Debiasing’☆59Updated 6 months ago
- ACL'24 (Oral) Tuning Large Multimodal Models for Videos using Reinforcement Learning from AI Feedback☆76Updated last year
- [ICCV 2025] VisRL: Intention-Driven Visual Perception via Reinforced Reasoning☆42Updated last month
- Task Preference Optimization: Improving Multimodal Large Language Models with Vision Task Alignment☆64Updated 5 months ago
- Official PyTorch code of GroundVQA (CVPR'24)☆64Updated last year
- [NeurIPS 2024] TransAgent: Transfer Vision-Language Foundation Models with Heterogeneous Agent Collaboration☆25Updated last year
- [ICLR 2025] Video-STaR: Self-Training Enables Video Instruction Tuning with Any Supervision☆72Updated last year
- [ECCV 2024] Learning Video Context as Interleaved Multimodal Sequences☆40Updated 9 months ago
- [NeurIPS 2024] Calibrated Self-Rewarding Vision Language Models☆83Updated 2 months ago
- ☆96Updated 6 months ago
- VideoHallucer, The first comprehensive benchmark for hallucination detection in large video-language models (LVLMs)☆41Updated 2 weeks ago
- Official repository of DoraemonGPT: Toward Understanding Dynamic Scenes with Large Language Models☆88Updated last year
- [ICLR 2025] AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmark☆137Updated 6 months ago
- VideoNIAH: A Flexible Synthetic Method for Benchmarking Video MLLMs☆53Updated 9 months ago
- [ICCV 2025] Official Repository of VideoLLaMB: Long Video Understanding with Recurrent Memory Bridges☆80Updated 10 months ago
- ✨✨The Curse of Multi-Modalities (CMM): Evaluating Hallucinations of Large Multimodal Models across Language, Visual, and Audio☆51Updated 5 months ago
- Code for "CAFe: Unifying Representation and Generation with Contrastive-Autoregressive Finetuning"☆29Updated 9 months ago