Sherrylone / PQDiffLinks
[ICLR 2024] Continuous-Multiple Image Outpainting in One-Step via Positional Query and A Diffusion-based Approach Link: https://arxiv.org/abs/2401.15652
β91Updated last month
Alternatives and similar repositories for PQDiff
Users that are interested in PQDiff are comparing it to the libraries listed below
Sorting:
- ποΈ Official implementation of "Gen4Gen: Generative Data Pipeline for Generative Multi-Concept Composition"β109Updated 2 months ago
- Implementation code of the paper MIGE: A Unified Framework for Multimodal Instruction-Based Image Generation and Editingβ72Updated 6 months ago
- [CVPR`2024, Oral] Attention Calibration for Disentangled Text-to-Image Personalizationβ109Updated last year
- [ACM Multimedia 2025 Datasets Track] EditWorld: Simulating World Dynamics for Instruction-Following Image Editingβ138Updated 6 months ago
- [ICLR 2024] Official PyTorch/Diffusers implementation of "Object-aware Inversion and Reassembly for Image Editing"β88Updated last year
- Official GitHub repository for the Text-Guided Video Editing (TGVE) competition of LOVEU Workshop @ CVPR'23.β78Updated 2 years ago
- Official Repo for Paper "OmniEdit: Building Image Editing Generalist Models Through Specialist Supervision" [ICLR2025]β141Updated last year
- "FreeU: Free Lunch in Diffusion U-Net" for Huggingface Diffusersβ102Updated 2 years ago
- Code for ICLR 2024 paper "Motion Guidance: Diffusion-Based Image Editing with Differentiable Motion Estimators"β108Updated last month
- (CVPR 2024) Official code for paper "Towards Language-Driven Video Inpainting via Multimodal Large Language Models"β99Updated last year
- we propose to generate a series of geometric shapes with target colors to disentangle (or peel off ) the target colors from the shapes. Bβ¦β69Updated last year
- [ICLR 2025] HQ-Edit: A High-Quality and High-Coverage Dataset for General Image Editingβ113Updated last year
- EditReward: A Human-Aligned Reward Model for Instruction-Guided Image Editing [ICLR 2026]β118Updated this week
- Code for the paper "Pix2Video: Video Editing using Image Diffusion"β76Updated 2 years ago
- β35Updated last year
- [ICLR2025] A versatile image-to-image visual assistant, designed for image generation, manipulation, and translation based on free-from uβ¦β210Updated 9 months ago
- β104Updated last year
- Implementation of paper EditCLIP: Representation Learning for Image Editing (ICCV 2025)β35Updated 7 months ago
- [NeurIPS 2023] Free-Bloom: Zero-Shot Text-to-Video Generator with LLM Director and LDM Animatorβ98Updated last year
- CVPR-24 | Official codebase for ZONE: Zero-shot InstructiON-guided Local Editingβ83Updated last year
- [CVPR 2024] Official repo for "InteractDiffusion: Interaction-Control for Text-to-Image Diffusion Model".β126Updated 7 months ago
- This is an official repository for the paper, NoiseCollage, which is a revolutionary extension of text-to-image diffusion models for layoβ¦β63Updated last year
- ICCV2023-Diffusion-Papersβ108Updated 2 years ago
- [ICLR 24] MaGIC: Multi-modality Guided Image Completionβ52Updated last year
- Official source codes of "TweedieMix: Improving Multi-Concept Fusion for Diffusion-based Image/Video Generation" (ICLR 2025)β62Updated last year
- [CVPR2024] Official code for Drag Your Noise: Interactive Point-based Editing via Diffusion Semantic Propagationβ87Updated last year
- CCEdit: Creative and Controllable Video Editing via Diffusion Modelsβ114Updated last year
- β109Updated last year
- [NeurIPS 2024] RealCompo: Balancing Realism and Compositionality Improves Text-to-Image Diffusion Modelsβ119Updated last year
- Official implementation of ImprovingText-guided ObjectInpainting with SemanticPre-inpainting in ECCV 2024β63Updated last year