HumanMLLM / HumanOmni
HumanOmni
β129Updated 2 weeks ago
Alternatives and similar repositories for HumanOmni:
Users that are interested in HumanOmni are comparing it to the libraries listed below
- This is the official implementation of "Flash-VStream: Memory-Based Real-Time Understanding for Long Video Streams"β173Updated 3 months ago
- The Next Step Forward in Multimodal LLM Alignmentβ135Updated 3 weeks ago
- π₯π₯First-ever hour scale video understanding modelsβ259Updated this week
- Long Context Transfer from Language to Visionβ368Updated last week
- A Unified Tokenizer for Visual Generation and Understandingβ210Updated 3 weeks ago
- β¨First Open-Source R1-like Video-LLM [2025/02/18]β289Updated last month
- β181Updated 8 months ago
- This is the official implementation of our paper "Video-RAG: Visually-aligned Retrieval-Augmented Long Video Comprehension"β161Updated last month
- A Simple Framework of Small-scale Large Multimodal Models for Video Understanding Based on TinyLLaVA_Factory.β46Updated last week
- Explore the Limits of Omni-modal Pretraining at Scaleβ97Updated 6 months ago
- β70Updated 2 weeks ago
- LinVT: Empower Your Image-level Large Language Model to Understand Videosβ64Updated 2 months ago
- [ICLR 2025] AuroraCap: Efficient, Performant Video Detailed Captioning and a New Benchmarkβ86Updated 2 months ago
- β40Updated last month
- SlowFast-LLaVA: A Strong Training-Free Baseline for Video Large Language Modelsβ206Updated 6 months ago
- LVBench: An Extreme Long Video Understanding Benchmarkβ85Updated 6 months ago
- Migician: Revealing the Magic of Free-Form Multi-Image Grounding in Multimodal Large Language Modelsβ47Updated 2 months ago
- [ACL 2024 Findings] "TempCompass: Do Video LLMs Really Understand Videos?", Yuanxin Liu, Shicheng Li, Yi Liu, Yuxiang Wang, Shuhuai Ren, β¦β107Updated last month
- Repository for 23'MM accepted paper "Curriculum-Listener: Consistency- and Complementarity-Aware Audio-Enhanced Temporal Sentence Groundiβ¦β49Updated last year
- Vision Search Assistant: Empower Vision-Language Models as Multimodal Search Enginesβ117Updated 4 months ago
- Official Repository of VideoLLaMB: Long Video Understanding with Recurrent Memory Bridgesβ66Updated last month
- [ECCV 2024] ShareGPT4V: Improving Large Multi-modal Models with Better Captionsβ210Updated 8 months ago
- β130Updated last month
- This is the official code of VideoAgent: A Memory-augmented Multimodal Agent for Video Understanding (ECCV 2024)β182Updated 3 months ago
- Tarsier -- a family of large-scale video-language models, which is designed to generate high-quality video descriptions , together with gβ¦β327Updated this week
- A jounery to real multimodel R1 ! We are doing on large-scale experimentβ280Updated 3 weeks ago
- β176Updated 8 months ago
- LongLLaVA: Scaling Multi-modal LLMs to 1000 Images Efficiently via Hybrid Architectureβ200Updated 2 months ago
- [CVPR 2025] VoCo-LLaMA: This repo is the official implementation of "VoCo-LLaMA: Towards Vision Compression with Large Language Models".β138Updated 3 weeks ago
- MM-EUREKA: Exploring Visual Aha Moment with Rule-based Large-scale Reinforcement Learningβ425Updated last week