ExponentialML / Text-To-Video-FinetuningLinks
Finetune ModelScope's Text To Video model using Diffusers π§¨
β688Updated last year
Alternatives and similar repositories for Text-To-Video-Finetuning
Users that are interested in Text-To-Video-Finetuning are comparing it to the libraries listed below
Sorting:
- [ICLR 2024] Official pytorch implementation of "ControlVideo: Training-free Controllable Text-to-Video Generation"β845Updated last year
- Official repo for VideoComposer: Compositional Video Synthesis with Motion Controllabilityβ944Updated last year
- Official Implementation of "Control-A-Video: Controllable Text-to-Video Generation with Diffusion Models"β398Updated 2 years ago
- [IJCV 2024] LaVie: High-Quality Video Generation with Cascaded Latent Diffusion Modelsβ935Updated 10 months ago
- [ICLR 2024] SEINE: Short-to-Long Video Diffusion Model for Generative Transition and Predictionβ942Updated 10 months ago
- Prompt-Free Diffusion: Taking "Text" out of Text-to-Image Diffusion Models, arxiv 2023 / CVPR 2024β757Updated last year
- [ECCV 2024] FreeInit: Bridging Initialization Gap in Video Diffusion Modelsβ529Updated last year
- [ICCV 2023] Consistent Image Synthesis and Editingβ811Updated last year
- Unoffical implement for [StyleDrop](https://arxiv.org/abs/2306.00983)β582Updated 2 years ago
- Transfer the ControlNet with any basemodel in diffusersπ₯β842Updated 2 years ago
- The official implementation for "Gen-L-Video: Multi-Text to Long Video Generation via Temporal Co-Denoising".β303Updated last year
- [CVPR 2024] PAIR Diffusion: A Comprehensive Multimodal Object-Level Image Editorβ519Updated last year
- β¨ Hotshot-XL: State-of-the-art AI text-to-GIF model trained to work alongside Stable Diffusion XLβ1,106Updated last year
- Video-P2P: Video Editing with Cross-attention Controlβ421Updated 3 months ago
- ICLR 2024 (Spotlight)β774Updated last year
- LVDM: Latent Video Diffusion Models for High-Fidelity Long Video Generationβ494Updated 10 months ago
- [NeurIPS 2023] Uni-ControlNet: All-in-One Control to Text-to-Image Diffusion Modelsβ655Updated last year
- ControlLoRA: A Lightweight Neural Network To Control Stable Diffusion Spatial Informationβ611Updated last year
- [IJCV] FastComposer: Tuning-Free Multi-Subject Image Generation with Localized Attentionβ708Updated 9 months ago
- Official Pytorch Implementation for "MultiDiffusion: Fusing Diffusion Paths for Controlled Image Generation" presenting "MultiDiffusion" β¦β1,042Updated 2 years ago
- Official implementation of "DreamPose: Fashion Image-to-Video Synthesis via Stable Diffusion"β1,008Updated last year
- ELITE: Encoding Visual Concepts into Textual Embeddings for Customized Text-to-Image Generation (ICCV 2023, Oral)β541Updated last year
- [ICCV 2023 Oral] "FateZero: Fusing Attentions for Zero-shot Text-based Video Editing"β1,149Updated 2 years ago
- Code for Enhancing Detail Preservation for Customized Text-to-Image Generation: A Regularization-Free Approachβ467Updated last year
- Official Pytorch Implementation for "Text2LIVE: Text-Driven Layered Image and Video Editing" (ECCV 2022 Oral)β890Updated 2 years ago
- Official PyTorch implementation for the paper "AnimateZero: Video Diffusion Models are Zero-Shot Image Animators"β351Updated last year
- β471Updated 3 months ago
- Code and data for "AnyV2V: A Tuning-Free Framework For Any Video-to-Video Editing Tasks" [TMLR 2024]β624Updated 11 months ago
- Make-A-Protagonist: Generic Video Editing with An Ensemble of Expertsβ323Updated 2 years ago
- MagicAvatar: Multimodal Avatar Generation and Animationβ622Updated 2 years ago