microsoft / CoMLLinks
Interactive coding assistant for data scientists and machine learning developers, empowered by large language models.
β95Updated 10 months ago
Alternatives and similar repositories for CoML
Users that are interested in CoML are comparing it to the libraries listed below
Sorting:
- π€ The code for "Can Large Language Model Agents Simulate Human Trust Behaviors?"β90Updated 4 months ago
- Open Implementations of LLM Analysesβ105Updated 10 months ago
- β59Updated 8 months ago
- Dedicated to building industrial foundation models for universal data intelligence across industries.β57Updated 11 months ago
- Formal-LLM: Integrating Formal Language and Natural Language for Controllable LLM-based Agentsβ125Updated last year
- Meta-CoT: Generalizable Chain-of-Thought Prompting in Mixed-task Scenarios with Large Language Modelsβ97Updated last year
- Official implementation of paper "On the Diagram of Thought" (https://arxiv.org/abs/2409.10038)β184Updated 4 months ago
- π§ Compare how Agent systems perform on several benchmarks. ππβ99Updated this week
- Resources for our paper: "EvoAgent: Towards Automatic Multi-Agent Generation via Evolutionary Algorithms"β120Updated 9 months ago
- The code implementation of MAGDi: Structured Distillation of Multi-Agent Interaction Graphs Improves Reasoning in Smaller Language Modelsβ¦β36Updated last year
- Codebase accompanying the Summary of a Haystack paper.β79Updated 10 months ago
- Lean implementation of various multi-agent LLM methods, including Iteration of Thought (IoT)β118Updated 5 months ago
- Beating the GAIA benchmark with Transformers Agents. πβ131Updated 5 months ago
- A re-implementation of Meta-Prompt in LangChain for building self-improving agents.β63Updated 2 years ago
- Code repo for MathAgentβ17Updated last year
- Official repo of Respond-and-Respond: data, code, and evaluationβ103Updated last year
- [ICLR 2024] Skeleton-of-Thought: Prompting LLMs for Efficient Parallel Generationβ174Updated last year
- A benchmark for evaluating learning agents based on just language feedbackβ86Updated last month
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing Systemβ131Updated last year
- Large Language Model (LLM) powered evaluator for Retrieval Augmented Generation (RAG) pipelines.β31Updated last year
- β66Updated 4 months ago
- β77Updated 10 months ago
- [ICLR 2025] A trinity of environments, tools, and benchmarks for general virtual agentsβ214Updated last month
- EcoAssistant: using LLM assistant more affordably and accuratelyβ132Updated last year
- Automatic Prompt Optimizationβ40Updated last year
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.β165Updated last week
- β123Updated 11 months ago
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Reasoning with Minimal Examplesβ103Updated 2 weeks ago
- [ACL25' Findings] SWE-Dev is an SWE agent with a scalable test case construction pipeline.β50Updated 2 weeks ago
- The Library for LLM-based multi-agent applicationsβ92Updated 3 weeks ago