Not-Diamond / awesome-ai-model-routing
A curated list of awesome approaches to AI model routing
☆94Updated this week
Alternatives and similar repositories for awesome-ai-model-routing:
Users that are interested in awesome-ai-model-routing are comparing it to the libraries listed below
- Routing on Random Forest (RoRF)☆136Updated 6 months ago
- AWM: Agent Workflow Memory☆252Updated last month
- Tutorial for building LLM router☆186Updated 8 months ago
- Sandboxed code execution for AI agents, locally or on the cloud. Massively parallel, easy to extend. Powering SWE-agent and more.☆129Updated last week
- ☆376Updated 2 months ago
- Research repository on interfacing LLMs with Weaviate APIs. Inspired by the Berkeley Gorilla LLM.☆119Updated last month
- Synthetic Data for LLM Fine-Tuning☆113Updated last year
- Code for ScribeAgent paper☆54Updated 3 weeks ago
- Code for Husky, an open-source language agent that solves complex, multi-step reasoning tasks. Husky v1 addresses numerical, tabular and …☆338Updated 9 months ago
- Testing and evaluation framework for voice agents☆98Updated last month
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing System☆109Updated 9 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆165Updated 3 weeks ago
- 🔧 Compare how Agent systems perform on several benchmarks. 📊🚀☆92Updated 5 months ago
- Code and Data for Tau-Bench☆358Updated 2 months ago
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆425Updated 6 months ago
- Task-based Agentic Framework using StrictJSON as the core☆451Updated last month
- Attribute (or cite) statements generated by LLMs back to in-context information.☆219Updated 5 months ago
- Function Calling Benchmark & Testing☆84Updated 8 months ago
- Scaling inference-time compute for LLM-as-a-judge, automated evaluations, guardrails, and reinforcement learning.☆189Updated last week
- ☆160Updated 7 months ago
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆104Updated 6 months ago
- ☆106Updated last week
- Beating the GAIA benchmark with Transformers Agents. 🚀☆103Updated last month
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆83Updated last week
- A simple unified framework for evaluating LLMs☆206Updated 2 weeks ago
- AgentLab: An open-source framework for developing, testing, and benchmarking web agents on diverse tasks, designed for scalability and re…☆283Updated last week
- Readymade evaluators for agent trajectories☆78Updated this week
- ☆143Updated this week
- ☆185Updated last month
- Resources for our paper: "Agent-R: Training Language Model Agents to Reflect via Iterative Self-Training"☆112Updated last week