AIGText / Glyph-ByT5
[ECCV2024] This is an official inference code of the paper "Glyph-ByT5: A Customized Text Encoder for Accurate Visual Text Rendering" and "Glyph-ByT5-v2: A Strong Aesthetic Baseline for Accurate Multilingual Visual Text Rendering""
☆481Updated 2 months ago
Related projects: ⓘ
- AutoStudio: Crafting Consistent Subjects in Multi-turn Interactive Image Generation☆391Updated last month
- ☆357Updated this week
- Code and data for "AnyV2V: A Tuning-Free Framework For Any Video-to-Video Editing Tasks"☆453Updated 3 weeks ago
- Implementation of UltraPixel: Advancing Ultra-High-Resolution Image Synthesis to New Peaks☆444Updated 2 weeks ago
- Stable-Hair: Real-World Hair Transfer via Diffusion Model☆330Updated last month
- [ECCV 2024] OMG: Occlusion-friendly Personalized Multi-concept Generation In Diffusion Models☆614Updated 2 months ago
- Code for DesignEdit☆300Updated last month
- [ECCV 2024] MOFA-Video: Controllable Image Animation via Generative Motion Field Adaptions in Frozen Image-to-Video Diffusion Model.☆576Updated last month
- SCEPTER is an open-source framework used for training, fine-tuning, and inference with generative models.☆320Updated last month
- SEED-Story: Multimodal Long Story Generation with Large Language Model☆692Updated last month
- ☆338Updated 3 months ago
- [ECCV 2024] HiDiffusion: Increases the resolution and speed of your diffusion model by only adding a single line of code!☆735Updated last month
- [CVPR 2024] FRESCO: Spatial-Temporal Correspondence for Zero-Shot Video Translation☆711Updated 3 months ago
- Diffree: Text-Guided Shape Free Object Inpainting with Diffusion Model☆207Updated last month
- Official Pytorch implementation of StreamV2V.☆429Updated this week
- [Siggraph Asia 2024] Follow-Your-Emoji: This repo is the official implementation of "Follow-Your-Emoji: Fine-Controllable and Expressive …☆295Updated last week
- Put Your Face Everywhere in Seconds.☆310Updated 9 months ago
- 🔥 CharacterFactory: Sampling Consistent Characters with GANs for Diffusion Models☆189Updated 2 months ago
- Official implementation of MotionClone: Training-Free Motion Cloning for Controllable Video Generation☆363Updated last month
- ☆301Updated 2 months ago
- 📺 An End-to-End Solution for High-Resolution and Long Video Generation Based on Transformer Diffusion☆1,155Updated 3 weeks ago
- Multimodal Models in Real World☆372Updated 2 months ago
- [ICML 2024] MagicPose(also known as MagicDance): Realistic Human Poses and Facial Expressions Retargeting with Identity-aware Diffusion☆674Updated 2 months ago
- [ECCV 2024] PowerPaint, a versatile image inpainting model that supports text-guided object inpainting, object removal, image outpainting…☆556Updated last week
- MotionFollower: Editing Video Motion via Lightweight Score-Guided Diffusion☆182Updated 3 months ago
- [ECCV 2024] FreeInit: Bridging Initialization Gap in Video Diffusion Models☆476Updated 8 months ago
- Official code for the paper "StreamMultiDiffusion: Real-Time Interactive Generation with Region-Based Semantic Control."☆522Updated 3 weeks ago
- Code for [CVPR 2024] VideoSwap: Customized Video Subject Swapping with Interactive Semantic Point Correspondence☆343Updated 5 months ago
- ELLA: Equip Diffusion Models with LLM for Enhanced Semantic Alignment☆1,048Updated 2 months ago
- Controllable video and image Generation, SVD, Animate Anyone, ControlNet, ControlNeXt, LoRA☆1,255Updated last week