River-Zhang / Awesome-FLUX-DiT
A collection of diffusion models based on FLUX/DiT for image/video generation, editing, reconstruction, inpainting .etc.
☆36Updated this week
Alternatives and similar repositories for Awesome-FLUX-DiT:
Users that are interested in Awesome-FLUX-DiT are comparing it to the libraries listed below
- CAR: Controllable AutoRegressive Modeling for Visual Generation☆107Updated 4 months ago
- T2V-CompBench: A Comprehensive Benchmark for Compositional Text-to-video Generation☆72Updated last week
- PyTorch implementation of DiffMoE, TC-DiT, EC-DiT and Dense DiT☆62Updated 2 weeks ago
- Official Repo for Paper "AnyEdit: Mastering Unified High-Quality Image Editing for Any Idea"☆65Updated this week
- This is the official implementation for ControlVAR.☆100Updated 3 months ago
- [ICLR 2025] Trajectory Attention For Fine-grained Video Motion Control☆61Updated last month
- Official source codes of "TweedieMix: Improving Multi-Concept Fusion for Diffusion-based Image/Video Generation" (ICLR 2025)☆40Updated 2 months ago
- [CVPR 2024] Official PyTorch implementation of FreeCustom: Tuning-Free Customized Image Generation for Multi-Concept Composition☆144Updated 2 months ago
- Official Implementation of VideoGen-of-Thought: Step-by-step generating multi-shot video with minimal manual intervention☆33Updated last week
- ☆47Updated 3 months ago
- [Arxiv'25] BlobCtrl: A Unified and Flexible Framework for Element-level Image Generation and Editing☆69Updated 2 weeks ago
- The official implementation of the paper titled "StableV2V: Stablizing Shape Consistency in Video-to-Video Editing".☆148Updated 3 months ago
- ☆27Updated 4 months ago
- World Simulator Assistant for Physics-Aware Text-to-Video Generation☆13Updated this week
- Official implementation of LiFT: Leveraging Human Feedback for Text-to-Video Model Alignment.☆70Updated last week
- [NeurIPS 2024] COVE: Unleashing the Diffusion Feature Correspondence for Consistent Video Editing☆23Updated 3 months ago
- Official implementation of "IFAdapter: Instance Feature Control for Grounded Text-to-Image Generation".☆54Updated 6 months ago
- [ICLR 2025] ControlAR: Controllable Image Generation with Autoregressive Models☆215Updated 2 months ago
- [Neurips 2024] Video Diffusion Models are Training-free Motion Interpreter and Controller☆34Updated last month
- ☆77Updated 10 months ago
- Implements VAR+CLIP for text-to-image (T2I) generation☆133Updated 2 months ago
- [NeurIPS 2024] The official implement of research paper "FreeLong : Training-Free Long Video Generation with SpectralBlend Temporal Atten…☆41Updated last month
- Official Implementation of VideoDPO☆76Updated 2 months ago
- [CVPR 2025] Open implementation of "RandAR"☆81Updated 2 weeks ago
- ☆57Updated 4 months ago
- Affordance-Aware Object Insertion via Mask-Aware Dual Diffusion☆38Updated last month
- Video Generation, Physical Commonsense, Semantic Adherence, VideoCon-Physics☆88Updated 2 weeks ago
- ☆80Updated 3 weeks ago
- [NeurIPS 2024] Motion Consistency Model: Accelerating Video Diffusion with Disentangled Motion-Appearance Distillation☆62Updated 5 months ago
- UniCombine: Unified Multi-Conditional Combination with Diffusion Transformer☆52Updated 2 weeks ago