web-arena-x / visualwebarena
VisualWebArena is a benchmark for multimodal agents.
☆211Updated last month
Related projects: ⓘ
- Official implementation for "You Only Look at Screens: Multimodal Chain-of-Action Agents" (Findings of ACL 2024)☆177Updated 2 months ago
- An Analytical Evaluation Board of Multi-turn LLM Agents☆227Updated 4 months ago
- [NeurIPS 2022] 🛒WebShop: Towards Scalable Real-World Web Interaction with Grounded Language Agents☆256Updated 2 weeks ago
- Code for the paper 🌳 Tree Search for Language Model Agents☆124Updated last month
- The model, data and code for the visual GUI Agent SeeClick☆182Updated 3 weeks ago
- FireAct: Toward Language Agent Fine-tuning☆242Updated 10 months ago
- SwiftSage: A Generative Agent with Fast and Slow Thinking for Complex Interactive Tasks☆239Updated this week
- ☆111Updated 3 months ago
- Code repo for "WebArena: A Realistic Web Environment for Building Autonomous Agents"☆681Updated last month
- ToolkenGPT: Augmenting Frozen Language Models with Massive Tools via Tool Embeddings - NeurIPS 2023 (oral)☆230Updated 5 months ago
- Official Implementation of Dynamic LLM-Agent Network: An LLM-agent Collaboration Framework with Agent Team Optimization☆96Updated 4 months ago
- Sotopia: an Open-ended Social Learning Environment (ICLR 2024 spotlight)☆146Updated this week
- Generative Judge for Evaluating Alignment☆208Updated 8 months ago
- [NeurIPS 2023 D&B] Code repository for InterCode benchmark https://arxiv.org/abs/2306.14898☆182Updated 4 months ago
- ☆242Updated 2 weeks ago
- This is the official repo for "PromptAgent: Strategic Planning with Language Models Enables Expert-level Prompt Optimization". PromptAgen…☆180Updated last month
- Benchmarking LLMs with Challenging Tasks from Real Users☆182Updated last month
- Official repo for paper DigiRL: Training In-The-Wild Device-Control Agents with Autonomous Reinforcement Learning.☆200Updated last month
- ICML 2024: Improving Factuality and Reasoning in Language Models through Multiagent Debate☆332Updated 11 months ago
- This repo contains evaluation code for the paper "MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for E…☆323Updated last week
- RewardBench: the first evaluation tool for reward models.☆352Updated last week
- An extensible benchmark for evaluating large language models on planning☆248Updated 3 months ago
- Codes for our paper "ChatEval: Towards Better LLM-based Evaluators through Multi-Agent Debate"☆221Updated 5 months ago
- Benchmarks, environments, and toolkits for general computer agents☆154Updated this week
- An implemtation of Everyting of Thoughts (XoT).☆114Updated 6 months ago
- Self-Alignment with Principle-Following Reward Models☆144Updated 6 months ago
- [ACL 2024] AUTOACT: Automatic Agent Learning from Scratch for QA via Self-Planning☆162Updated 5 months ago
- Chain-of-Hindsight, A Scalable RLHF Method☆213Updated 11 months ago
- Towards Large Multimodal Models as Visual Foundation Agents☆87Updated 3 weeks ago
- A codebase for "Language Models can Solve Computer Tasks"☆218Updated 4 months ago