ZJU-REAL / Awesome-GUI-AgentsLinks
A curated collection of resources, tools, and frameworks for developing GUI Agents.
☆108Updated this week
Alternatives and similar repositories for Awesome-GUI-Agents
Users that are interested in Awesome-GUI-Agents are comparing it to the libraries listed below
Sorting:
- A collection of multimodal reasoning papers, codes, datasets, benchmarks and resources.☆279Updated 3 weeks ago
- Reverse Chain-of-Thought Problem Generation for Geometric Reasoning in Large Multimodal Models☆177Updated 9 months ago
- Chain-of-Spot: Interactive Reasoning Improves Large Vision-language Models☆97Updated last year
- A Gaussian dense reward framework for GUI grounding training☆204Updated 2 weeks ago
- [ICLR 2025] Mathematical Visual Instruction Tuning for Multi-modal Large Language Models☆148Updated 8 months ago
- (ECCV 2024) Empowering Multimodal Large Language Model as a Powerful Data Generator☆112Updated 4 months ago
- [Arxiv] Discrete Diffusion in Large Language and Multimodal Models: A Survey☆185Updated 3 weeks ago
- (ICCV 2025) Enhance CLIP and MLLM's fine-grained visual representations with generative models.☆68Updated last month
- Your efficient and accurate answer verification system for RL training.☆34Updated last month
- [ICCV 2025] SparseMM: Head Sparsity Emerges from Visual Concept Responses in MLLMs☆69Updated last month
- [NAACL 2025 Oral] 🎉 From redundancy to relevance: Enhancing explainability in multimodal large language models☆107Updated 5 months ago
- A collection of token reduction (token pruning, merging, clustering, etc.) techniques for ML/AI☆116Updated 2 weeks ago
- 🚀 [NeurIPS24] Make Vision Matter in Visual-Question-Answering (VQA)! Introducing NaturalBench, a vision-centric VQA benchmark (NeurIPS'2…☆84Updated last month
- ✨✨Long-VITA: Scaling Large Multi-modal Models to 1 Million Tokens with Leading Short-Context Accuracy☆292Updated 2 months ago
- [ECCV 2024] Does Your Multi-modal LLM Truly See the Diagrams in Visual Math Problems?☆167Updated 3 months ago
- Recipes to train the self-rewarding reasoning LLMs.☆224Updated 5 months ago
- [ECCV 2024] Efficient Inference of Vision Instruction-Following Models with Elastic Cache☆42Updated last year
- A scalable, end-to-end training pipeline for general-purpose agents☆349Updated last month
- ✨✨R1-Reward: Training Multimodal Reward Model Through Stable Reinforcement Learning☆246Updated 2 months ago
- [ICML 2025] Official repository for paper "Scaling Video-Language Models to 10K Frames via Hierarchical Differential Distillation"☆165Updated 2 months ago
- Efficient Reasoning Vision Language Models☆337Updated 2 weeks ago
- A library for generating difficulty-scalable, multi-tool, and verifiable agentic tasks with execution trajectories.☆141Updated 3 weeks ago
- An open-source implementation for training LLaVA-NeXT.☆413Updated 9 months ago
- This is the repo for the paper "OS Agents: A Survey on MLLM-based Agents for Computer, Phone and Browser Use" (ACL 2025 Oral).☆313Updated last month
- GPT-ImgEval: Evaluating GPT-4o’s state-of-the-art image generation capabilities☆286Updated 3 months ago
- WorldGPT: Empowering LLM as Multimodal World Model☆117Updated last year
- Explore concepts like Self-Correct, Self-Refine, Self-Improve, Self-Contradict, Self-Play, and Self-Knowledge, alongside o1-like reasonin…☆170Updated 7 months ago
- Official code of paper "Beyond 'Aha!': Toward Systematic Meta-Abilities Alignment in Large Reasoning Models"☆79Updated 2 months ago
- The code for "TokenPacker: Efficient Visual Projector for Multimodal LLM", IJCV2025☆263Updated 2 months ago
- [ICLR 2025] MLLM for On-Demand Spatial-Temporal Understanding at Arbitrary Resolution☆318Updated last month