ali-vilab / composerLinks
Official implementation of "Composer: Creative and Controllable Image Synthesis with Composable Conditions"
☆1,562Updated last year
Alternatives and similar repositories for composer
Users that are interested in composer are comparing it to the libraries listed below
Sorting:
- Official Pytorch Implementation for "MultiDiffusion: Fusing Diffusion Paths for Controlled Image Generation" presenting "MultiDiffusion" …☆1,041Updated last year
- Versatile Diffusion: Text, Images and Variations All in One Diffusion Model, arXiv 2022 / ICCV 2023☆1,332Updated last year
- Unofficial implementation of "Prompt-to-Prompt Image Editing with Cross Attention Control" with Stable Diffusion☆1,336Updated 2 years ago
- Prompt-Free Diffusion: Taking "Text" out of Text-to-Image Diffusion Models, arxiv 2023 / CVPR 2024☆751Updated last year
- Official Pytorch Implementation for "Text2LIVE: Text-Driven Layered Image and Video Editing" (ECCV 2022 Oral)☆889Updated 2 years ago
- Transfer the ControlNet with any basemodel in diffusers🔥☆832Updated 2 years ago
- Zero-shot Image-to-Image Translation [SIGGRAPH 2023]☆1,124Updated 8 months ago
- Unoffical implement for [StyleDrop](https://arxiv.org/abs/2306.00983)☆583Updated last year
- Official Pytorch Implementation for “Plug-and-Play Diffusion Features for Text-Driven Image-to-Image Translation” (CVPR 2023)☆980Updated 2 years ago
- [ICML'23] StyleGAN-T: Unlocking the Power of GANs for Fast Large-Scale Text-to-Image Synthesis☆1,183Updated 2 years ago
- Open-Set Grounded Text-to-Image Generation☆2,133Updated last year
- ControlLoRA: A Lightweight Neural Network To Control Stable Diffusion Spatial Information☆601Updated 11 months ago
- Custom Diffusion: Multi-Concept Customization of Text-to-Image Diffusion (CVPR 2023)☆1,948Updated last year
- Consistency Distilled Diff VAE☆2,190Updated last year
- Rich-Text-to-Image Generation☆792Updated last year
- Implementation of Phenaki Video, which uses Mask GIT to produce text guided videos of up to 2 minutes in length, in Pytorch☆776Updated 11 months ago
- ☆3,342Updated last year
- ☆1,476Updated last year
- Official implementation of "DreamPose: Fashion Image-to-Video Synthesis via Stable Diffusion"☆1,008Updated last year
- Unified Controllable Visual Generation Model☆649Updated 5 months ago
- A large-scale text-to-image prompt gallery dataset based on Stable Diffusion☆1,288Updated last year
- [IJCV] FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attention☆705Updated 6 months ago
- T2I-Adapter☆3,708Updated last year
- [ICCV 2023 Oral] "FateZero: Fusing Attentions for Zero-shot Text-based Video Editing"☆1,148Updated last year
- [ICLR 2024] Official pytorch implementation of "ControlVideo: Training-free Controllable Text-to-Video Generation"☆834Updated last year
- Implementation of Paint-with-words with Stable Diffusion : method from eDiff-I that let you generate image from text-labeled segmentation…☆645Updated 2 years ago
- Implementation of GigaGAN, new SOTA GAN out of Adobe. Culmination of nearly a decade of research into GANs☆1,911Updated 5 months ago
- Official Implementation for "Attend-and-Excite: Attention-Based Semantic Guidance for Text-to-Image Diffusion Models" (SIGGRAPH 2023)☆740Updated last year
- Official repo for VideoComposer: Compositional Video Synthesis with Motion Controllability☆937Updated last year
- Code for Enhancing Detail Preservation for Customized Text-to-Image Generation: A Regularization-Free Approach☆466Updated last year