feizc / Video-Stable-Diffusion
Generate consistent videos with stable diffusion models
β48Updated 2 years ago
Alternatives and similar repositories for Video-Stable-Diffusion:
Users that are interested in Video-Stable-Diffusion are comparing it to the libraries listed below
- A one-stop library to standardize the inference and evaluation of all the conditional video generation models.β47Updated 2 weeks ago
- π€ Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch - fork with video pseudo3dβ98Updated last year
- Implementation of Text2Video-Zero: Text-to-Image Diffusion Models are Zero-Shot Video Generatorsβ85Updated last year
- Implementation of the text to video model LUMIERE from the paper: "A Space-Time Diffusion Model for Video Generation" by Google Researchβ51Updated this week
- My implementation of the model KosmosG from "KOSMOS-G: Generating Images in Context with Multimodal Large Language Models"β14Updated 2 months ago
- β3Updated 4 months ago
- [ICLR 2024] Code for FreeNoise based on AnimateDiffβ106Updated last year
- Official implementation of "VSTAR: Generative Temporal Nursing for Longer Dynamic Video Synthesis"β19Updated this week
- Official implementation for "pOps: Photo-Inspired Diffusion Operators"β75Updated 6 months ago
- Extend BoxDiff to SDXL (SDXL-based layout-to-image generation)β20Updated 8 months ago
- β73Updated last year
- Code Release for the paper "Make-A-Story: Visual Memory Conditioned Consistent Story Generation" in CVPR 2023β37Updated last year
- β35Updated 9 months ago
- ποΈ Official implementation of "Gen4Gen: Generative Data Pipeline for Generative Multi-Concept Composition"β104Updated 8 months ago
- Implementation of Collage Diffusion (https://arxiv.org/abs/2303.00262)β35Updated last year
- An attempt at a SVD inpainting pipelineβ51Updated last year
- Ground-A-Video: Zero-shot Grounded Video Editing using Text-to-image Diffusion Models (ICLR 2024)β135Updated 8 months ago
- A retrain of AnimateDiff to be conditional on an init imageβ33Updated last year
- Blending Custom Photos with Video Diffusion Transformersβ40Updated last week
- β20Updated 4 months ago
- T2VScore: Towards A Better Metric for Text-to-Video Generationβ78Updated 9 months ago
- β65Updated last year
- Videoshop: Localized Semantic Video Editing with Noise-Extrapolated Diffusion Inversionβ37Updated 5 months ago
- π€ Unofficial huggingface/diffusers-based implementation of the paper "Training-Free Layout Control with Cross-Attention Guidance".β41Updated last year
- Reuse and Diffuse: Iterative Denoising for Text-to-Video Generationβ38Updated last year
- [TMLR] Official PyTorch implementation of "Ξ»-ECLIPSE: Multi-Concept Personalized Text-to-Image Diffusion Models by Leveraging CLIP Latentβ¦β51Updated 2 months ago
- We introduce OpenStory++, a large-scale open-domain dataset focusing on enabling MLLMs to perform storytelling generation tasks.β13Updated 5 months ago
- [WACV 2025] Follow-Your-Handle: This repo is the official implementation of "MagicStick: Controllable Video Editing via Control Handle Trβ¦β87Updated last year
- DiffBlender: Scalable and Composable Multimodal Text-to-Image Diffusion Modelsβ44Updated last year
- β92Updated last year