toto222 / DICE-TalkLinks
DICE-Talk is a diffusion-based emotional talking head generation method that can generate vivid and diverse emotions for speaking portraits.
☆290Updated 5 months ago
Alternatives and similar repositories for DICE-Talk
Users that are interested in DICE-Talk are comparing it to the libraries listed below
Sorting:
- [ICCV 2025] Official Pytorch Implementation of FLOAT: Generative Motion Latent Flow Matching for Audio-driven Talking Portrait.☆449Updated 2 months ago
- KeySync: A Robust Approach for Leakage-free Lip Synchronization in High Resolution☆374Updated 3 weeks ago
- 开源的LstmSync数字人泛化模型,只做最好的泛化模型!☆135Updated this week
- [AAAI 2026] EchoMimicV3: 1.3B Parameters are All You Need for Unified Multi-Modal and Multi-Task Human Animation☆697Updated last month
- A 2D customized lip-sync model for high-fidelity real-time driving.☆119Updated 6 months ago
- [CVPR-2025] The official code of HunyuanPortrait: Implicit Condition Control for Enhanced Portrait Animation☆327Updated last month
- [ECCV 2024 Oral] EDTalk - Official PyTorch Implementation☆455Updated 3 months ago
- [ACM MM 2025] Ditto: Motion-Space Diffusion for Controllable Realtime Talking Head Synthesis☆660Updated 2 months ago
- FantasyPortrait: Enhancing Multi-Character Portrait Animation with Expression-Augmented Diffusion Transformers☆492Updated 4 months ago
- SkyReels-A1: Expressive Portrait Animation in Video Diffusion Transformers☆575Updated 7 months ago
- [ICLR2025] DisPose: Disentangling Pose Guidance for Controllable Human Image Animation☆375Updated last month
- ☆81Updated 5 months ago
- project page for ChatAnyone☆116Updated 9 months ago
- ☆407Updated 6 months ago
- MagicTryOn is a video virtual try-on framework based on a large-scale video diffusion Transformer.☆495Updated 2 weeks ago
- LLIA - Enabling Low-Latency Interactive Avatars: Real-Time Audio-Driven Portrait Video Generation with Diffusion Models☆145Updated 7 months ago
- ICCV 2025 ACTalker: an end-to-end video diffusion framework for talking head synthesis that supports both single and multi-signal control…☆440Updated 4 months ago
- SoulX-FlashTalk is the first 14B model to achieve sub-second start-up latency (0.87s) while maintaining a real-time throughput of 32 FPS …☆193Updated this week
- [CVPR 2025 Highlight] X-Dyna: Expressive Dynamic Human Image Animation☆260Updated 11 months ago
- [ICCV2025] MotionFollower: Editing Video Motion via Lightweight Score-Guided Diffusion☆242Updated 6 months ago
- ☆239Updated last week
- Offical implement of Dynamic Frame Avatar with Non-autoregressive Diffusion Framework for talking head Video Generation☆237Updated 2 months ago
- Full version of wav2lip-onnx including face alignment and face enhancement and more...☆150Updated 7 months ago
- [NeurIPS 2025] OmniTalker: Real-Time Text-Driven Talking Head Generation with In-Context Audio-Visual Style Replication☆413Updated 3 months ago
- JoyHallo: Digital human model for Mandarin☆521Updated 3 months ago
- [ICLR 2025] Animate-X: Universal Character Image Animation with Enhanced Motion Representation☆379Updated 3 months ago
- This is the official implementation of our paper: "MiniMax-Remover: Taming Bad Noise Helps Video Object Removal"☆518Updated 5 months ago
- talking-face video editing☆413Updated 10 months ago
- [CVPR-2025] The official code of HunyuanPortrait: Implicit Condition Control for Enhanced Portrait Animation☆282Updated 7 months ago
- ☆648Updated last month