showlab / Show-Anything-3DLinks
Edit and Generate Anything in 3D world!
☆13Updated 2 years ago
Alternatives and similar repositories for Show-Anything-3D
Users that are interested in Show-Anything-3D are comparing it to the libraries listed below
Sorting:
- An interactive demo based on Segment-Anything for stroke-based painting which enables human-like painting.☆35Updated 2 years ago
- A curated list of papers and resources for text-to-image evaluation.☆29Updated last year
- Awesome-DragGAN: A curated list of papers, tutorials, repositories related to DragGAN☆85Updated last year
- Official repo for the TMLR paper "Discffusion: Discriminative Diffusion Models as Few-shot Vision and Language Learners"☆29Updated last year
- Code for Evolutionary Caching to Accelerate Your Off-the-Shelf Diffusion Model☆25Updated 3 weeks ago
- Description and applications of OpenAI's paper about DALL-E (2021) and implementation of other (CLIP-guided) zero-shot text-to-image gene…☆33Updated 2 years ago
- Accepted by AAAI2022☆21Updated 3 years ago
- ☆24Updated 2 years ago
- ☆10Updated last year
- ☆14Updated 9 months ago
- My implementation of the model KosmosG from "KOSMOS-G: Generating Images in Context with Multimodal Large Language Models"☆14Updated 8 months ago
- Motion-conditional image animation for video editing☆20Updated last year
- A benchmark dataset and simple code examples for measuring the perception and reasoning of multi-sensor Vision Language models.☆18Updated 6 months ago
- Democratising RGBA Image Generation With No $$$ (AI4VA@ECCV24)☆30Updated 10 months ago
- Code and Data for Paper: SELMA: Learning and Merging Skill-Specific Text-to-Image Experts with Auto-Generated Data☆34Updated last year
- [ECCV 2024] This is the official implementation of "Stitched ViTs are Flexible Vision Backbones".☆27Updated last year
- INF-LLaVA: Dual-perspective Perception for High-Resolution Multimodal Large Language Model☆42Updated 11 months ago
- This repository is associated with the research paper titled ImageChain: Advancing Sequential Image-to-Text Reasoning in Multimodal Large…☆12Updated last month
- This is the implementation of CounterCurate, the data curation pipeline of both physical and semantic counterfactual image-caption pairs.☆18Updated last year
- DDS: Delta Denoising Score PyTorch implementation☆19Updated last year
- VPEval Codebase from Visual Programming for Text-to-Image Generation and Evaluation (NeurIPS 2023)☆45Updated last year
- Code for paper <Explain Me the Painting: Multi-Topic Knowledgeable Art Description Generation> in ICCV 2021.☆13Updated 3 years ago
- Self-Supervised Object Detection via Generative Image Synthesis☆28Updated 3 years ago
- ☆14Updated 4 months ago
- ☆30Updated last year
- Grounding Language Models for Compositional and Spatial Reasoning☆17Updated 2 years ago
- Pytorch implementation of HyperLLaVA: Dynamic Visual and Language Expert Tuning for Multimodal Large Language Models☆28Updated last year
- This is a repository for my work on the paper "Oracle Guided Image Synthesis with Relative Queries".☆24Updated 3 years ago
- Open source community's implementation of the model from "LANGUAGE MODEL BEATS DIFFUSION — TOKENIZER IS KEY TO VISUAL GENERATION"☆15Updated 8 months ago
- Official InfiniBench: A Benchmark for Large Multi-Modal Models in Long-Form Movies and TV Shows☆15Updated last month