Yushi-Hu / VisualSketchpad
Codes for Visual Sketchpad: Sketching as a Visual Chain of Thought for Multimodal Language Models
☆160Updated 3 months ago
Alternatives and similar repositories for VisualSketchpad:
Users that are interested in VisualSketchpad are comparing it to the libraries listed below
- The official repository for "2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining"☆140Updated 3 weeks ago
- Insight-V: Exploring Long-Chain Visual Reasoning with Multimodal Large Language Models☆128Updated last month
- Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs☆73Updated 3 months ago
- Code for the paper "AutoPresent: Designing Structured Visuals From Scratch"☆48Updated last month
- [TMLR] Public code repo for paper "A Single Transformer for Scalable Vision-Language Modeling"☆127Updated 3 months ago
- Official code for Paper "Mantis: Multi-Image Instruction Tuning" (TMLR2024)☆197Updated this week
- CuMo: Scaling Multimodal LLM with Co-Upcycled Mixture-of-Experts☆139Updated 8 months ago
- [NeurIPS 2024] A task generation and model evaluation system for multimodal language models.☆63Updated 2 months ago
- Auto Interpretation Pipeline and many other functionalities for Multimodal SAE Analysis.☆104Updated 3 weeks ago
- Enhancing Large Vision Language Models with Self-Training on Image Comprehension.☆63Updated 8 months ago
- Code and data for OS-Genesis: Automating GUI Agent Trajectory Construction via Reverse Task Synthesis☆94Updated 3 weeks ago
- Rethinking Step-by-step Visual Reasoning in LLMs☆240Updated 3 weeks ago
- (ICLR 2025) The Official Code Repository for GUI-World.☆46Updated last month
- ☆134Updated 8 months ago
- A Survey on Benchmarks of Multimodal Large Language Models☆84Updated last month
- [ICLR 2025] Video-STaR: Self-Training Enables Video Instruction Tuning with Any Supervision☆59Updated 7 months ago
- ☆45Updated last month
- Official code of *Virgo: A Preliminary Exploration on Reproducing o1-like MLLM*☆86Updated last month
- An LLM-free Multi-dimensional Benchmark for Multi-modal Hallucination Evaluation☆111Updated last year
- Code for Math-LLaVA: Bootstrapping Mathematical Reasoning for Multimodal Large Language Models☆76Updated 7 months ago
- GUICourse: From General Vision Langauge Models to Versatile GUI Agents☆98Updated 6 months ago
- MATH-Vision dataset and code to measure Multimodal Mathematical Reasoning capabilities.☆85Updated 4 months ago
- [CVPR 2024] Prompt Highlighter: Interactive Control for Multi-Modal LLMs☆138Updated 6 months ago
- [IEEE VIS 2024] LLaVA-Chart: Advancing Multimodal Large Language Models in Chart Question Answering with Visualization-Referenced Instruc…☆61Updated 3 weeks ago
- This is the official repository of our paper "What If We Recaption Billions of Web Images with LLaMA-3 ?"☆128Updated 8 months ago
- Evaluation framework for paper "VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?"☆48Updated 3 months ago
- What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆58Updated 3 months ago
- Official repo for StableLLAVA☆94Updated last year
- [ICLR'25 Oral] UGround: Universal GUI Visual Grounding for GUI Agents☆160Updated this week
- LongLLaVA: Scaling Multi-modal LLMs to 1000 Images Efficiently via Hybrid Architecture☆189Updated last month