alibaba-damo-academy / LumosLinks
Lumos Project: Frontier generative model research by Alibaba DAMO Academy, including Lumos-1, etc.
☆80Updated this week
Alternatives and similar repositories for Lumos
Users that are interested in Lumos are comparing it to the libraries listed below
Sorting:
- Official PyTorch implementation - Video Motion Transfer with Diffusion Transformers☆67Updated 2 months ago
- CVPRW 2025 paper Progressive Autoregressive Video Diffusion Models: https://arxiv.org/abs/2410.08151☆76Updated 2 months ago
- Vision as a Dialect: Unifying Visual Understanding and Generation via Text-Aligned Representations☆106Updated last week
- Code for FreeTraj, a tuning-free method for trajectory-controllable video generation☆104Updated 11 months ago
- [ NeurIPS 2024 D&B Track ] Implementation for "FiVA: Fine-grained Visual Attribute Dataset for Text-to-Image Diffusion Models"☆70Updated 6 months ago
- ☆44Updated 2 months ago
- Subjects200K dataset☆114Updated 6 months ago
- [Arxiv'25] BlobCtrl: A Unified and Flexible Framework for Element-level Image Generation and Editing☆9Updated 3 months ago
- [Neurips 2024] Video Diffusion Models are Training-free Motion Interpreter and Controller☆44Updated 3 months ago
- [AAAI-2025] Official implementation of Image Conductor: Precision Control for Interactive Video Synthesis☆93Updated 11 months ago
- Official source codes of "TweedieMix: Improving Multi-Concept Fusion for Diffusion-based Image/Video Generation" (ICLR 2025)☆51Updated 5 months ago
- [NeurIPS 2024] The official implement of research paper "FreeLong : Training-Free Long Video Generation with SpectralBlend Temporal Atten…☆51Updated 2 weeks ago
- [NeurIPS 2024 Spotlight] The official implement of research paper "MotionBooth: Motion-Aware Customized Text-to-Video Generation"☆134Updated 9 months ago
- UniCombine: Unified Multi-Conditional Combination with Diffusion Transformer☆94Updated 2 weeks ago
- [ICCV 2025] MagicMotion: Controllable Video Generation with Dense-to-Sparse Trajectory Guidance☆137Updated 3 weeks ago
- I Think, Therefore I Diffuse: Enabling Multimodal In-Context Reasoning in Diffusion Models☆154Updated 4 months ago
- ☆33Updated 9 months ago
- Implementation code of the paper MIGE: A Unified Framework for Multimodal Instruction-Based Image Generation and Editing☆65Updated this week
- Official implementation of our paper: "Ca2-VDM: Efficient Autoregressive Video Diffusion Model with Causal Generation and Cache Sharing" …☆63Updated last month
- The official repository of "Sekai: A Video Dataset towards World Exploration"☆98Updated last week
- Official implementation of LiFT: Leveraging Human Feedback for Text-to-Video Model Alignment.☆79Updated 2 months ago
- GoT-R1: Unleashing Reasoning Capability of MLLM for Visual Generation with Reinforcement Learning☆89Updated last month
- [CVPR2025 Highlight] PAR: Parallelized Autoregressive Visual Generation. https://yuqingwang1029.github.io/PAR-project☆166Updated 3 months ago
- PyTorch implementation of DiffMoE, TC-DiT, EC-DiT and Dense DiT☆120Updated 2 months ago
- Code for ICLR 2024 paper "Motion Guidance: Diffusion-Based Image Editing with Differentiable Motion Estimators"☆103Updated last year
- Official implementation of MARS: Mixture of Auto-Regressive Models for Fine-grained Text-to-image Synthesis☆85Updated last year
- Reflect-DiT: Inference-Time Scaling for Text-to-Image Diffusion Transformers via In-Context Reflection☆42Updated 3 weeks ago
- Diffusion Powers Video Tokenizer for Comprehension and Generation (CVPR 2025)☆72Updated 4 months ago
- [ICCV 2025] Balanced Image Stylization with Style Matching Score☆54Updated 2 weeks ago
- Concat-ID: Towards Universal Identity-Preserving Video Synthesis☆54Updated 2 months ago