zzxslp / MM-NavigatorLinks
GPT-4V in Wonderland: LMMs as Smartphone Agents
☆135Updated last year
Alternatives and similar repositories for MM-Navigator
Users that are interested in MM-Navigator are comparing it to the libraries listed below
Sorting:
- Official implementation for "You Only Look at Screens: Multimodal Chain-of-Action Agents" (Findings of ACL 2024)☆255Updated last year
- ☆66Updated 2 years ago
- [ICLR'25 Oral] UGround: Universal GUI Visual Grounding for GUI Agents☆294Updated 6 months ago
- ControlLLM: Augment Language Models with Tools by Searching on Graphs☆194Updated last year
- GUICourse: From General Vision Langauge Models to Versatile GUI Agents☆135Updated last year
- WebLINX is a benchmark for building web navigation agents with conversational capabilities☆157Updated 11 months ago
- Code for Paper: Harnessing Webpage Uis For Text Rich Visual Understanding☆53Updated last year
- A Universal Platform for Training and Evaluation of Mobile Interaction☆60Updated 4 months ago
- [ICLR 2025] A trinity of environments, tools, and benchmarks for general virtual agents☆221Updated 7 months ago
- ☆59Updated 2 years ago
- (ICLR 2025) The Official Code Repository for GUI-World.☆67Updated last year
- Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMs☆99Updated last year
- ☆35Updated 2 years ago
- [ICLR 2024] Trajectory-as-Exemplar Prompting with Memory for Computer Control☆65Updated 2 weeks ago
- [NeurIPS 2024 D&B] VideoGUI: A Benchmark for GUI Automation from Instructional Videos☆48Updated 7 months ago
- Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters☆91Updated 2 years ago
- The model, data and code for the visual GUI Agent SeeClick☆460Updated 6 months ago
- Code for Paper: Autonomous Evaluation and Refinement of Digital Agents [COLM 2024]☆148Updated last year
- [ECCV2024] 🐙Octopus, an embodied vision-language model trained with RLEF, emerging superior in embodied visual planning and programming.☆294Updated last year
- Towards Large Multimodal Models as Visual Foundation Agents☆252Updated 9 months ago
- ScreenQA dataset was introduced in the "ScreenQA: Large-Scale Question-Answer Pairs over Mobile App Screenshots" paper. It contains ~86K …☆139Updated 11 months ago
- Evaluation framework for paper "VisualWebBench: How Far Have Multimodal LLMs Evolved in Web Page Understanding and Grounding?"☆63Updated last year
- ☆20Updated last year
- Code repo for "Read Anywhere Pointed: Layout-aware GUI Screen Reading with Tree-of-Lens Grounding"☆28Updated last year
- The Screen Annotation dataset consists of pairs of mobile screenshots and their annotations. The annotations are in text format, and desc…☆84Updated last year
- Recent advancements propelled by large language models (LLMs), encompassing an array of domains including Vision, Audio, Agent, Robotics,…☆124Updated 7 months ago
- [ACL 2024] PCA-Bench: Evaluating Multimodal Large Language Models in Perception-Cognition-Action Chain☆106Updated last year
- ☆123Updated last year
- Official repo for paper DigiRL: Training In-The-Wild Device-Control Agents with Autonomous Reinforcement Learning.☆387Updated 11 months ago
- [NeurIPS 2025 Spotlight] Scaling Computer-Use Grounding via UI Decomposition and Synthesis☆144Updated 2 months ago