Yaofang-Liu / Mochi-Full-FinetunerLinks
Code for full fintuing Mochi model with FSDP (and CP)
☆27Updated 3 months ago
Alternatives and similar repositories for Mochi-Full-Finetuner
Users that are interested in Mochi-Full-Finetuner are comparing it to the libraries listed below
Sorting:
- ☆67Updated last year
- [ACM MM24] Official implementation of ACM MM 2024 paper: "ZePo: Zero-Shot Portrait Stylization with Faster Sampling"☆41Updated 10 months ago
- ☆4Updated 9 months ago
- ☆62Updated last year
- Official code of "LayerTracer: Cognitive-Aligned Layered SVG Synthesis via Diffusion Transformer"☆55Updated 3 months ago
- This is the official repository for "LatentMan: Generating Consistent Animated Characters using Image Diffusion Models" [CVPRW 2024]☆23Updated 11 months ago
- Omegance: A Single Parameter for Various Granularities in Diffusion-Based Synthesis (ICCV, 2025)☆52Updated 2 weeks ago
- AAAI 2025: Anywhere: A Multi-Agent Framework for User-Guided, Reliable, and Diverse Foreground-Conditioned Image Generation☆41Updated last year
- ☆45Updated last month
- The public source code of "FreCaS: Efficient Higher-Resolution Image Generation via Frequency-aware Cascaded Sampling"☆28Updated last week
- Collection of scripts to build small-scale datasets for fine-tuning video generation models.☆63Updated 3 months ago
- TextBoost: Towards One-Shot Personalization of Text-to-Image Models via Fine-tuning Text Encoder☆57Updated 5 months ago
- [ECCV 2024] IDOL: Unified Dual-Modal Latent Diffusion for Human-Centric Joint Video-Depth Generation☆55Updated 9 months ago
- CutDiffusion: A Simple, Fast, Cheap, and Strong Diffusion Extrapolation Method☆27Updated last year
- Blending Custom Photos with Video Diffusion Transformers☆47Updated 5 months ago
- [ICLR 2024] Code for FreeNoise based on AnimateDiff☆107Updated last year
- [ECCVW 2024] Prompt Sliders for Fine-Grained Control, Editing and Erasing of Concepts in Diffusion Models☆29Updated 2 months ago
- ☆34Updated 6 months ago
- [ICLR 2024] Code for FreeNoise based on LaVie☆34Updated last year
- [WACV 2025] MegaFusion: Extend Diffusion Models towards Higher-resolution Image Generation without Further Tuning☆92Updated 2 months ago
- This respository contains the code for the NeurIPS 2024 paper SF-V: Single Forward Video Generation Model.☆97Updated 7 months ago
- ☆20Updated 9 months ago
- Concat-ID: Towards Universal Identity-Preserving Video Synthesis☆54Updated 2 months ago
- [ICLR 2025] Official lmplementation of SPM-Diff: Incorporating Visual Correspondence into Diffusion Model for Virtual Try-On☆38Updated 4 months ago
- FantasyID: Face Knowledge Enhanced ID-Preserving Video Generation☆65Updated 2 months ago
- ☆85Updated 10 months ago
- DMM: Building a Versatile Image Generation Model via Distillation-Based Model Merging☆45Updated 2 months ago
- [ICML 2025] Official PyTorch implementation of paper "Ultra-Resolution Adaptation with Ease".☆99Updated 2 months ago
- Public code release for the paper "ProCreate, Don’t Reproduce! Propulsive Energy Diffusion for Creative Generation"☆40Updated 2 months ago
- DiT for VAE (and Video Generation)☆34Updated 10 months ago