sagekit / webvoyagerLinks
Magnitude achieves SOTA 94% on WebVoyager benchmark
☆29Updated 4 months ago
Alternatives and similar repositories for webvoyager
Users that are interested in webvoyager are comparing it to the libraries listed below
Sorting:
- The fastest, lightest, and easiest-to-integrate AI gateway on the market. Fully open-sourced.☆462Updated 3 months ago
- Prompt engineering, automated.☆348Updated 7 months ago
- Infrastructure that's powering E2B Cloud.☆727Updated this week
- Multi-language code navigation API in a container☆95Updated 3 months ago
- A fully customizable and self-hosted sandboxing solution for AI agent code execution and computer use. It features out-of-the-box support…☆660Updated 5 months ago
- A simple Python sandbox for helpful LLM data agents☆294Updated last year
- Python SDK for running evaluations on LLM generated responses☆293Updated 5 months ago
- 🤖 Headless IDE for AI agents☆200Updated last month
- Kura is a simple reproduction of the CLIO paper which uses language models to label user behaviour before clustering them based on embedd…☆366Updated 2 months ago
- Inference-time scaling for LLMs-as-a-judge.☆310Updated 2 weeks ago
- Sidecar is the AI brains for the Aide editor and works alongside it, locally on your machine☆589Updated 6 months ago
- Open source AI Agent evaluation framework for web tasks 🐒🍌☆312Updated 10 months ago
- ACP is the Agent Control Plane - a distributed agent scheduler optimized for simplicity, clarity, and control. It is designed for outer-l…☆246Updated 4 months ago
- Giving Claude ability to run code with E2B via MCP (Model Context Protocol)☆349Updated 2 weeks ago
- Memory Library for Building Agents with Social Intelligence☆243Updated this week
- Routing on Random Forest (RoRF)☆219Updated last year
- 📚 Benchmark your browser agent on ~2.5k READ and ACTION based tasks☆73Updated 3 months ago
- A tool kit for generating high quality prompts using DSPy GEPA optimizer☆285Updated last month
- Deep Research for your internal data☆348Updated 5 months ago
- A toolkit for building computer use AI agents☆178Updated 4 months ago
- AutoEvals is a tool for quickly and easily evaluating AI model outputs using best practices.☆722Updated this week
- Open-source versioning, tracing, and annotation tooling.☆204Updated 2 weeks ago
- ☆142Updated 8 months ago
- Provider-agnostic, open-source evaluation infrastructure for language models☆653Updated last week
- ☆478Updated this week
- Work with web-enabled agents quickly — whether running a quick task or bootstrapping a full-stack product.☆93Updated last year
- An MCP Server that's also an MCP Client. Useful for letting Claude develop and test MCPs without needing to reset the application.☆124Updated 8 months ago
- The easiest, and fastest way to run AI-generated Python code safely☆339Updated 11 months ago
- Postman for MCP servers☆123Updated 3 months ago
- Testing and evaluation framework for voice agents☆154Updated 5 months ago