JosephPai / Art-Description
Code for paper <Explain Me the Painting: Multi-Topic Knowledgeable Art Description Generation> in ICCV 2021.
☆13Updated 3 years ago
Related projects: ⓘ
- Repository for the data in the paper "Explain Me the Painting: Multi-TopicKnowledgeable Art Description Generation".☆17Updated 3 years ago
- CVPR2023 paper☆50Updated last year
- ☆52Updated 4 months ago
- PEA-Diffusion: Parameter-Efficient Adapter with Knowledge Distillation in non-English Text-to-Image Generation☆19Updated 2 months ago
- Code release for our paper "Divide and Conquer: Language Models can Plan and Self-Correct for Compositional Text-to-Image Generation".☆17Updated 7 months ago
- Simple script to compute CLIP-based scores given a DALL-e trained model.☆30Updated 3 years ago
- VPEval Codebase from Visual Programming for Text-to-Image Generation and Evaluation (NeurIPS 2023)☆42Updated 9 months ago
- ☆46Updated last year
- Code and Data for Paper: SELMA: Learning and Merging Skill-Specific Text-to-Image Experts with Auto-Generated Data☆30Updated 6 months ago
- Official implementation and dataset for the NAACL 2024 paper "ComCLIP: Training-Free Compositional Image and Text Matching"☆30Updated last month
- INF-LLaVA: Dual-perspective Perception for High-Resolution Multimodal Large Language Model☆36Updated last month
- Official PyTorch implementation of "λ-ECLIPSE: Multi-Concept Personalized Text-to-Image Diffusion Models by Leveraging CLIP Latent Space"☆43Updated 5 months ago
- Code Release for the paper "Make-A-Story: Visual Memory Conditioned Consistent Story Generation" in CVPR 2023☆37Updated last year
- ☆34Updated last month
- A Unified Framework for Video-Language Understanding☆55Updated last year
- Code for CVPR'2022 paper ✨ "Predict, Prevent, and Evaluate: Disentangled Text-Driven Image Manipulation Empowered by Pre-Trained Vision-L…☆37Updated 2 years ago
- Research code for "Training Vision-Language Transformers from Captions Alone"☆34Updated 2 years ago
- Visual Instruction-guided Explainable Metric. Code for "Towards Explainable Metrics for Conditional Image Synthesis Evaluation" (ACL 2024…☆22Updated last month
- [ECCV 2022] FashionViL: Fashion-Focused V+L Representation Learning☆56Updated last year
- Official repo for the paper "Discffusion: Discriminative Diffusion Models as Few-shot Vision and Language Learners"☆26Updated 4 months ago
- Using LLMs and pre-trained caption models for super-human performance on image captioning.☆40Updated 11 months ago
- [CVPR 2023] VoP: Text-Video Co-operative Prompt Tuning for Cross-Modal Retrieval☆36Updated last year
- Video shot transition detection☆21Updated last year
- Masked Vision-Language Transformer in Fashion☆32Updated 11 months ago
- This is the official repository for the paper "OpenFashionCLIP: Vision-and-Language Contrastive Learning with Open-Source Fashion Data". …☆52Updated 4 months ago
- FuseCap: Large Language Model for Visual Data Fusion in Enriched Caption Generation☆47Updated 5 months ago
- kdexd/coco-caption@de6f385☆26Updated 4 years ago
- Codes and Models for COSA: Concatenated Sample Pretrained Vision-Language Foundation Model☆38Updated last year
- CLIP4IDC: CLIP for Image Difference Captioning (AACL 2022)☆27Updated last year
- Official Pytorch implementation of "CompoDiff: Versatile Composed Image Retrieval With Latent Diffusion" (TMLR 2024)☆73Updated last month