MeiGen-AI / InfiniteTalkLinks
Unlimited-length talking video generation that supports image-to-video and video-to-video generation
☆3,527Updated 3 months ago
Alternatives and similar repositories for InfiniteTalk
Users that are interested in InfiniteTalk are comparing it to the libraries listed below
Sorting:
- [NeurIPS 2025] Let Them Talk: Audio-Driven Multi-Person Conversational Video Generation☆2,683Updated 2 months ago
- [ACM MM 2025] FantasyTalking: Realistic Talking Portrait Generation via Coherent Motion Synthesis☆1,592Updated 3 months ago
- ☆1,946Updated last month
- SkyReels V1: The first and most advanced open-source human-centric video foundation model☆2,489Updated 8 months ago
- Taming Stable Diffusion for Lip Sync!☆5,177Updated 5 months ago
- [NeurIPS 2025] OmniTalker: Real-Time Text-Driven Talking Head Generation with In-Context Audio-Visual Style Replication☆397Updated 2 months ago
- SkyReels-V2: Infinite-length Film Generative model☆5,052Updated 3 months ago
- HunyuanCustom: A Multimodal-Driven Architecture for Customized Video Generation☆1,193Updated last month
- Official implementation of "Sonic: Shifting Focus to Global Audio Perception in Portrait Animation"☆3,122Updated 5 months ago
- Sonic is a method about ' Shifting Focus to Global Audio Perception in Portrait Animation',you can use it in comfyUI☆1,111Updated 2 months ago
- ☆2,788Updated last week
- VoxCPM: Tokenizer-Free TTS for Context-Aware Speech Generation and True-to-Life Voice Cloning☆2,198Updated last month
- [CVPR 2025] EchoMimicV2: Towards Striking, Simplified, and Semi-Body Human Animation☆4,380Updated 3 months ago
- [AAAI 2026] EchoMimicV3: 1.3B Parameters are All You Need for Unified Multi-Modal and Multi-Task Human Animation☆638Updated last week
- This node provides lip-sync capabilities in ComfyUI using ByteDance's LatentSync model. It allows you to synchronize video lips with audi…☆919Updated 2 months ago
- SoulX-Podcast is an inference codebase by the Soul AI team for generating high-fidelity podcasts from text.☆2,340Updated this week
- Phantom: Subject-Consistent Video Generation via Cross-Modal Alignment☆1,455Updated 2 months ago
- An Open-Source Multimodal AIGC Solution based on ComfyUI + MCP + LLM https://pixelle.ai☆794Updated 3 weeks ago
- Diffusion-based Portrait and Animal Animation☆844Updated 2 months ago
- [NeurIPS 2025] Image editing is worth a single LoRA! 0.1% training data for fantastic image editing! Surpasses GPT-4o in ID persistence~ …☆2,037Updated 2 weeks ago
- SkyReels-A1: Expressive Portrait Animation in Video Diffusion Transformers☆569Updated 5 months ago
- ☆636Updated last week
- [SIGGRAPH 2025] LAM: Large Avatar Model for One-shot Animatable Gaussian Head☆849Updated 2 months ago
- [AAAI 2025] EchoMimic: Lifelike Audio-Driven Portrait Animations through Editable Landmark Conditioning☆4,117Updated 3 months ago
- HuMo: Human-Centric Video Generation via Collaborative Multi-Modal Conditioning☆928Updated last month
- [CVPR 2025] MMAudio: Taming Multimodal Joint Training for High-Quality Video-to-Audio Synthesis☆1,979Updated 2 months ago
- AutoClip : AI-powered video clipping and highlight generation · 一款智能高光提取与剪辑的二创工具☆960Updated 2 months ago
- ☆5,348Updated last week
- 一个超轻量级、可以在移动端实时运行的数字人模型☆2,330Updated 2 months ago
- AutoClip: AI-powered video clipping and highlight generation · 一款智能高光提取与剪辑的二创工具☆928Updated 2 months ago