InternLM / StarBenchLinks
☆34Updated last month
Alternatives and similar repositories for StarBench
Users that are interested in StarBench are comparing it to the libraries listed below
Sorting:
- Kling-Foley: Multimodal Diffusion Transformer for High-Quality Video-to-Audio Generation☆61Updated 5 months ago
- [ECCV 2024 Oral] Audio-Synchronized Visual Animation☆58Updated last year
- DeepDubber-V1: Towards High Quality and Dialogue, Narration, Monologue Adaptive Movie Dubbing Via Multi-Modal Chain-of-Thoughts Reasoning…☆26Updated 3 months ago
- [CVPR 2024] Seeing and Hearing: Open-domain Visual-Audio Generation with Diffusion Latent Aligners☆153Updated last year
- [ICCV2025] TokenBridge: Bridging Continuous and Discrete Tokens for Autoregressive Visual Generation. https://yuqingwang1029.github.io/To…☆150Updated 4 months ago
- The author's implementation of FUDOKI, a multimodal large language model purely based on discrete flow matching.☆63Updated 2 months ago
- ☆37Updated 3 months ago
- [Arxiv 2024] Official code for MMTrail: A Multimodal Trailer Video Dataset with Language and Music Descriptions☆33Updated 10 months ago
- ACDiT: Interpolating Autoregressive Conditional Modeling and Diffusion Transformer☆38Updated 11 months ago
- video-SALMONN 2 is a powerful audio-visual large language model (LLM) that generates high-quality audio-visual video captions, which is d…☆123Updated last month
- ☆185Updated 11 months ago
- [🏆 IJCV 2025 & ACCV 2024 Best Paper Honorable Mention] Official pytorch implementation of the paper "High-Quality Visually-Guided Sound …☆23Updated last month
- [AAAI 2025] Empowering LLMs with Pseudo-Untrimmed Videos for Audio-Visual Temporal Understanding☆33Updated 8 months ago
- Official PyTorch implementation of EMOVA in CVPR 2025 (https://arxiv.org/abs/2409.18042)☆74Updated 8 months ago
- Diff-Foley: Synchronized Video-to-Audio Synthesis with Latent Diffusion Models☆198Updated last year
- Official source codes for the paper: EmoDubber: Towards High Quality and Emotion Controllable Movie Dubbing.☆32Updated 6 months ago
- LongVALE: Vision-Audio-Language-Event Benchmark Towards Time-Aware Omni-Modal Perception of Long Videos. (CVPR 2025))☆54Updated 6 months ago
- official code for CVPR'24 paper Diff-BGM☆71Updated last year
- The official implementation of OmniFlow: Any-to-Any Generation with Multi-Modal Rectified Flows☆119Updated 3 months ago
- Official Repository of IJCAI 2024 Paper: "BATON: Aligning Text-to-Audio Model with Human Preference Feedback"☆30Updated 9 months ago
- Tracking the latest and greatest research papers on video generation.☆97Updated this week
- ☆10Updated 8 months ago
- a fully open-source implementation of a GPT-4o-like speech-to-speech video understanding model.☆35Updated 8 months ago
- ☆105Updated 6 months ago
- [NeurIPS 2025] HermesFlow: Seamlessly Closing the Gap in Multimodal Understanding and Generation☆73Updated 2 months ago
- [ICCV 2025] Official repo for "GigaTok: Scaling Visual Tokenizers to 3 Billion Parameters for Autoregressive Image Generation"☆194Updated 5 months ago
- ☆62Updated 5 months ago
- Towards Fine-grained Audio Captioning with Multimodal Contextual Cues☆84Updated 2 months ago
- EchoInk-R1: Exploring Audio-Visual Reasoning in Multimodal LLMs via Reinforcement Learning [🔥The Exploration of R1 for General Audio-Vi…☆68Updated 6 months ago
- Official Code for "ARM-Thinker: Reinforcing Multimodal Generative Reward Models with Agentic Tool Use and Visual Reasoning"☆40Updated this week