apple / ToolSandboxLinks
☆236Updated 2 months ago
Alternatives and similar repositories for ToolSandbox
Users that are interested in ToolSandbox are comparing it to the libraries listed below
Sorting:
- Complex Function Calling Benchmark.☆163Updated last year
- Self-Reflection in LLM Agents: Effects on Problem-Solving Performance☆93Updated last year
- ☆242Updated last year
- AWM: Agent Workflow Memory☆387Updated last month
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆120Updated 3 months ago
- xLAM: A Family of Large Action Models to Empower AI Agent Systems☆599Updated 5 months ago
- 🌍 AppWorld: A Controllable World of Apps and People for Benchmarking Function Calling and Interactive Coding Agent, ACL'24 Best Resource…☆367Updated 2 months ago
- LOFT: A 1 Million+ Token Long-Context Benchmark☆225Updated 7 months ago
- Code for the paper 🌳 Tree Search for Language Model Agents☆219Updated last year
- Beating the GAIA benchmark with Transformers Agents. 🚀☆145Updated 11 months ago
- Comprehensive benchmark for RAG☆260Updated 7 months ago
- MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents [EMNLP 2024]☆194Updated 5 months ago
- Meta Agents Research Environments is a comprehensive platform designed to evaluate AI agents in dynamic, realistic scenarios. Unlike stat…☆418Updated last week
- An Analytical Evaluation Board of Multi-turn LLM Agents [NeurIPS 2024 Oral]☆389Updated last year
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing System☆152Updated last year
- Official Implementation of "Multi-Head RAG: Solving Multi-Aspect Problems with LLMs"☆237Updated 4 months ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuning☆366Updated last year
- WorkBench: a Benchmark Dataset for Agents in a Realistic Workplace Setting.☆61Updated last month
- Official repo for "LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs".☆241Updated last year
- ToolBench, an evaluation suite for LLM tool manipulation capabilities.☆172Updated last year
- Official implementation of paper "On the Diagram of Thought" (https://arxiv.org/abs/2409.10038)☆193Updated last week
- ☆328Updated 6 months ago
- Official repository for paper "ReasonIR Training Retrievers for Reasoning Tasks".☆217Updated 7 months ago
- (ACL 2025 Main) Code for MultiAgentBench : Evaluating the Collaboration and Competition of LLM agents https://www.arxiv.org/pdf/2503.019…☆213Updated 3 months ago
- 🔧 Compare how Agent systems perform on several benchmarks. 📊🚀☆103Updated 6 months ago
- ☆217Updated this week
- A banchmark list for evaluation of large language models.☆159Updated 2 weeks ago
- [ACL 2024] AutoAct: Automatic Agent Learning from Scratch for QA via Self-Planning☆233Updated last year
- ToolQA, a new dataset to evaluate the capabilities of LLMs in answering challenging questions with external tools. It offers two levels …☆285Updated 2 years ago
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"☆218Updated last year