atfortes / Awesome-Controllable-Diffusion
Papers and resources on Controllable Generation using Diffusion Models, including ControlNet, DreamBooth, IP-Adapter.
β467Updated last month
Alternatives and similar repositories for Awesome-Controllable-Diffusion:
Users that are interested in Awesome-Controllable-Diffusion are comparing it to the libraries listed below
- π₯π₯π₯ A curated list of papers on LLMs-based multimodal generation (image, video, 3D and audio).β471Updated last month
- π This is a repository for organizing papers, codes and other resources related to unified multimodal models.β535Updated last month
- [ICLR 2024 Spotlight] DreamLLM: Synergistic Multimodal Comprehension and Creationβ438Updated 5 months ago
- LaVIT: Empower the Large Language Model to Understand and Generate Visual Contentβ578Updated 7 months ago
- Official implementation of SEED-LLaMA (ICLR 2024).β612Updated 7 months ago
- A collection of resources on controllable generation with text-to-image diffusion models.β1,034Updated 4 months ago
- Paper list about multimodal and large language models, only used to record papers I read in the daily arxiv for personal needs.β621Updated this week
- Chatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing imagβ¦β519Updated last year
- π Code and models for the NeurIPS 2023 paper "Generating Images with Multimodal Language Models".β457Updated last year
- (CVPR2024)A benchmark for evaluating Multimodal LLMs using multiple-choice questions.β338Updated 3 months ago
- PyTorch implementation of InstructDiffusion, a unifying and generic framework for aligning computer vision tasks with human instructions.β427Updated 11 months ago
- Research Trends in LLM-guided Multimodal Learning.β358Updated last year
- Aligning LMMs with Factually Augmented RLHFβ361Updated last year
- A list of works on evaluation of visual generation models, including evaluation metrics, models, and systemsβ284Updated 2 weeks ago
- π A curated list of resources dedicated to hallucination of multimodal large language models (MLLM).β673Updated last month
- [Survey] Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Surveyβ422Updated 3 months ago
- LLM-grounded Diffusion: Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models (LLM-grounded Diffusiβ¦β470Updated 7 months ago
- The official GitHub page for the review paper "Sora: A Review on Background, Technology, Limitations, and Opportunities of Large Vision Mβ¦β498Updated last year
- Recent LLM-based CV and related works. Welcome to comment/contribute!β861Updated 2 months ago
- β¨β¨Woodpecker: Hallucination Correction for Multimodal Large Language Modelsβ634Updated 4 months ago
- A reading list of video generationβ563Updated last week
- Awesome_Multimodel is a curated GitHub repository that provides a comprehensive collection of resources for Multimodal Large Language Modβ¦β321Updated last month
- Diffusion Model-Based Image Editing: A Survey (TPAMI 2025)β607Updated last month
- [Neurips'24 Spotlight] Visual CoT: Advancing Multi-Modal Language Models with a Comprehensive Dataset and Benchmark for Chain-of-Thought β¦β307Updated 4 months ago
- β328Updated last year
- Official PyTorch implementation of the paper "In-Context Learning Unlocked for Diffusion Models"β404Updated last year
- Multimodal Models in Real Worldβ502Updated 2 months ago
- [ICLR'24] Mitigating Hallucination in Large Multi-Modal Models via Robust Instruction Tuningβ277Updated last year
- [TMLR 2025π₯] A survey for the autoregressive models in vision.β542Updated last week
- [ICLR 2025] Repository for Show-o, One Single Transformer to Unify Multimodal Understanding and Generation.β1,380Updated last week