google-research-datasets / uicritLinks
UICrit is a dataset containing human-generated natural language design critiques, corresponding bounding boxes for each critique, and design quality ratings for 1,000 mobile UIs from RICO. This dataset was collected for our UIST '24 paper: https://arxiv.org/abs/2407.08850.
☆22Updated 9 months ago
Alternatives and similar repositories for uicrit
Users that are interested in uicrit are comparing it to the libraries listed below
Sorting:
- Code for the paper "AutoPresent: Designing Structured Visuals From Scratch" (CVPR 2025)☆120Updated 3 months ago
- VPEval Codebase from Visual Programming for Text-to-Image Generation and Evaluation (NeurIPS 2023)☆45Updated last year
- Code repo for "Read Anywhere Pointed: Layout-aware GUI Screen Reading with Tree-of-Lens Grounding"☆28Updated last year
- Official Repo of Graphist☆125Updated last year
- Official implementation of "Describing Differences in Image Sets with Natural Language" (CVPR 2024 Oral)☆120Updated last year
- Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs☆89Updated 10 months ago
- Continuous diffusion for layout generation☆47Updated 6 months ago
- [CVPR 2023 highlight] Towards Flexible Multi-modal Document Models☆59Updated last year
- Official repository for LLaVA-Reward (ICCV 2025): Multimodal LLMs as Customized Reward Models for Text-to-Image Generation☆20Updated last month
- CycleReward is a reward model trained on cycle consistency preferences to measure image-text alignment.☆38Updated 3 weeks ago
- Code base of SynthCLIP: CLIP training with purely synthetic text-image pairs from LLMs and TTIs.☆100Updated 5 months ago
- ☆17Updated 11 months ago
- Official code repo for "Editing Implicit Assumptions in Text-to-Image Diffusion Models"☆86Updated 2 years ago
- ☆25Updated last year
- Code for the paper "If at First You Don't Succeed, Try, Try Again: Faithful Diffusion-based Text-to-Image Generation by Selection"☆27Updated 2 years ago
- ☆37Updated last year
- Code and data for EMNLP 2023 paper "Grounding Visual Illusions in Language: Do Vision-Language Models Perceive Illusions Like Humans?"☆14Updated last year
- [NeurIPS2023] Official implementation and model release of the paper "What Makes Good Examples for Visual In-Context Learning?"☆177Updated last year
- OpenCOLE: Towards Reproducible Automatic Graphic Design Generation [Inoue+, CVPRW2024 (GDUG)]☆78Updated 5 months ago
- ECCV2024_Parrot Captions Teach CLIP to Spot Text☆66Updated 11 months ago
- ☆41Updated last year
- VINS: Visual Search for Mobile User Interface Design☆44Updated 4 years ago
- [NeurIPS 2024] Efficient Large Multi-modal Models via Visual Context Compression☆62Updated 6 months ago
- ☆21Updated last year
- [ICLR 2024] Official code for the paper "LLM Blueprint: Enabling Text-to-Image Generation with Complex and Detailed Prompts"☆81Updated last year
- Implementation and dataset for paper "Can MLLMs Perform Text-to-Image In-Context Learning?"☆41Updated 2 months ago
- COLA: Evaluate how well your vision-language model can Compose Objects Localized with Attributes!☆24Updated 9 months ago
- ☆70Updated 2 months ago
- Code, Data and Red Teaming for ZeroBench☆46Updated 3 months ago
- VideoRFT: Incentivizing Video Reasoning Capability in MLLMs via Reinforced Fine-Tuning☆36Updated this week