OmniMMI / OpenOmniNexus
a fully open-source implementation of a GPT-4o-like speech-to-speech video understanding model.
☆9Updated 2 weeks ago
Alternatives and similar repositories for OpenOmniNexus:
Users that are interested in OpenOmniNexus are comparing it to the libraries listed below
- OpenOmni: Official implementation of Advancing Open-Source Omnimodal Large Language Models with Progressive Multimodal Alignment and Rea…☆41Updated last month
- [CVPR 2025] OmniMMI: A Comprehensive Multi-modal Interaction Benchmark in Streaming Video Contexts☆11Updated 2 weeks ago
- [ACL2023] VSTAR is a multimodal dialogue dataset with scene and topic transition information☆12Updated 5 months ago
- A project for tri-modal LLM benchmarking and instruction tuning.☆32Updated 3 weeks ago
- ☆11Updated 2 months ago
- LMM solved catastrophic forgetting, AAAI2025☆40Updated last week
- ☆18Updated 11 months ago
- Code for Talk With Human-like Agents: Empathetic Dialogue Through Perceptible Acoustic Reception and Reaction (ACL24))☆42Updated 8 months ago
- ☆38Updated 8 months ago
- ☆20Updated 10 months ago
- UnifiedMLLM: Enabling Unified Representation for Multi-modal Multi-tasks With Large Language Model☆22Updated 8 months ago
- Narrative movie understanding benchmark☆70Updated 11 months ago
- BLSP-Emo: Towards Empathetic Large Speech-Language Models☆43Updated 10 months ago
- The code and weight for LoVA. LoVA is a novel model for Long-form Video-to-Audio generation. Based on the Diffusion Transformer (DiT) arc…☆13Updated last month
- Multimodal Empathetic Chatbot☆37Updated 9 months ago
- ☆21Updated last year
- [ECCV’24] Official Implementation for CAT: Enhancing Multimodal Large Language Model to Answer Questions in Dynamic Audio-Visual Scenario…☆52Updated 7 months ago
- Pytorch Implementation of the Model from "MIRASOL3B: A MULTIMODAL AUTOREGRESSIVE MODEL FOR TIME-ALIGNED AND CONTEXTUAL MODALITIES"☆26Updated 2 months ago
- WorldSense: Evaluating Real-world Omnimodal Understanding for Multimodal LLMs☆22Updated this week
- A collection of omni-mllm☆21Updated last week
- [NAACL 2024] LaDiC: Are Diffusion Models Really Inferior to Autoregressive Counterparts for Image-to-text Generation?☆38Updated 10 months ago
- [2024-ACL]: TextBind: Multi-turn Interleaved Multimodal Instruction-following in the Wildrounded Conversation☆47Updated last year
- ☆55Updated 9 months ago
- [ACL 2024] A Multimodal, Multigenre, and Multipurpose Audio-Visual Academic Lecture Dataset☆14Updated last week
- [ICLR 2025] CREMA: Generalizable and Efficient Video-Language Reasoning via Multimodal Modular Fusion☆41Updated 3 months ago
- Official PyTorch implementation of EMOVA in CVPR 2025 (https://arxiv.org/abs/2409.18042)☆29Updated last month
- LUCY: Linguistic Understanding and Control Yielding Early Stage of Her☆37Updated last week
- ☆26Updated 3 weeks ago
- Official Repository of IJCAI 2024 Paper: "BATON: Aligning Text-to-Audio Model with Human Preference Feedback"☆27Updated last month
- The dataset and baseline code for Text-to-Audio Grounding (TAG)☆42Updated 3 months ago