yhZhai / idolLinks
[ECCV 2024] IDOL: Unified Dual-Modal Latent Diffusion for Human-Centric Joint Video-Depth Generation
☆55Updated last year
Alternatives and similar repositories for idol
Users that are interested in idol are comparing it to the libraries listed below
Sorting:
- [arXiv'25] AnyCharV: Bootstrap Controllable Character Video Generation with Fine-to-Coarse Guidance☆40Updated 9 months ago
- This is the official repository for "LatentMan: Generating Consistent Animated Characters using Image Diffusion Models" [CVPRW 2024]☆21Updated last year
- HyperMotion is a pose guided human image animation framework based on a large-scale video diffusion Transformer.☆124Updated 4 months ago
- [ACM MM24] MotionMaster: Training-free Camera Motion Transfer For Video Generation☆97Updated last year
- MotionShop: Zero-Shot Motion Transfer in Video Diffusion Models with Mixture of Score Guidance☆26Updated 11 months ago
- ☆43Updated last year
- Experiencing lightning fast (~1s) and accurate drag-based image editing☆82Updated last year
- Official code for VividPose: Advancing Stable Video Diffusion for Realistic Human Image Animation.☆84Updated last year
- ☆32Updated 8 months ago
- ☆66Updated last year
- Official Repository of [CVPR'24 Highlight Diffportrait3D: Controllable Diffusion for Zero-Shot Portrait View Synthesis]☆148Updated 10 months ago
- This is the project for 'Any2Caption', Interpreting Any Condition to Caption for Controllable Video Generation☆45Updated 7 months ago
- [ECCV 2024] Noise Calibration: Plug-and-play Content-Preserving Video Enhancement using Pre-trained Video Diffusion Models☆87Updated last year
- ☆50Updated last month
- [ICLR 2024] Code for FreeNoise based on LaVie☆33Updated last year
- [CVPR 2025] Official code for "Synergizing Motion and Appearance: Multi-Scale Compensatory Codebooks for Talking Head Video Generation"☆64Updated 5 months ago
- We introduce OpenStory++, a large-scale open-domain dataset focusing on enabling MLLMs to perform storytelling generation tasks.☆15Updated last year
- [ICLR 2024] Code for FreeNoise based on AnimateDiff☆108Updated last year
- AniCrafter: Customizing Realistic Human-Centric Animation via Avatar-Background Conditioning in Video Diffusion Models☆126Updated 4 months ago
- Official Pytorch Implementation of Paper - Stable-Pose: Leveraging Transformers for Pose-Guided Text-to-Image Generation - NeurIPS 2024☆110Updated 11 months ago
- [CVPR 2024] Make-It-Vivid: Dressing Your Animatable Biped Cartoon Characters from Text☆71Updated last year
- This respository contains the code for the NeurIPS 2024 paper SF-V: Single Forward Video Generation Model.☆99Updated 11 months ago
- AAAI 2025: Anywhere: A Multi-Agent Framework for User-Guided, Reliable, and Diverse Foreground-Conditioned Image Generation☆44Updated last year
- ☆65Updated last year
- Official implementation of "Slicedit: Zero-Shot Video Editing With Text-to-Image Diffusion Models Using Spatio-Temporal Slices" (ICML 202…☆58Updated 11 months ago
- Official pytorch implementation for SingleInsert☆27Updated last year
- [ICCV 2025] Pytorch implementation of "PersonaCraft: Personalized Full-Body Image Synthesis for Multiple Identities from Single Reference…☆51Updated 8 months ago
- [SIGGRAPH Asia 2024] TrailBlazer: Trajectory Control for Diffusion-Based Video Generation☆99Updated last year
- [AAAI'25] Official implementation of Image Conductor: Precision Control for Interactive Video Synthesis☆99Updated last year
- [CVPR 2025] Zero-1-to-A: Zero-Shot One Image to Animatable Head Avatars Using Video Diffusion☆42Updated 8 months ago