microsoft / CoML
Interactive coding assistant for data scientists and machine learning developers, empowered by large language models.
☆93Updated 7 months ago
Alternatives and similar repositories for CoML
Users that are interested in CoML are comparing it to the libraries listed below
Sorting:
- DSBench: How Far are Data Science Agents from Becoming Data Science Experts?☆52Updated 2 months ago
- Codebase accompanying the Summary of a Haystack paper.☆78Updated 7 months ago
- Lean implementation of various multi-agent LLM methods, including Iteration of Thought (IoT)☆110Updated 3 months ago
- 🤝 The code for "Can Large Language Model Agents Simulate Human Trust Behaviors?"☆81Updated last month
- Code for MultiAgentBench : Evaluating the Collaboration and Competition of LLM agents https://www.arxiv.org/pdf/2503.01935☆100Updated last month
- Dedicated to building industrial foundation models for universal data intelligence across industries.☆53Updated 8 months ago
- Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆90Updated 2 months ago
- ☆56Updated 5 months ago
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.☆142Updated this week
- ☆47Updated 4 months ago
- [ICLR 2024] Skeleton-of-Thought: Prompting LLMs for Efficient Parallel Generation☆167Updated last year
- Benchmark and research code for the paper SWEET-RL Training Multi-Turn LLM Agents onCollaborative Reasoning Tasks☆193Updated last week
- Resources for our paper: "EvoAgent: Towards Automatic Multi-Agent Generation via Evolutionary Algorithms"☆97Updated 6 months ago
- Open Implementations of LLM Analyses☆102Updated 7 months ago
- Verifiers for LLM Reinforcement Learning☆50Updated last month
- Self-Reflection in LLM Agents: Effects on Problem-Solving Performance☆73Updated 5 months ago
- [ICML 2025] Flow of Reasoning: Training LLMs for Divergent Problem Solving with Minimal Examples☆85Updated last month
- Scalable Meta-Evaluation of LLMs as Evaluators☆42Updated last year
- Deep Research through Multi-Agents, using GraphRAG☆69Updated 6 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆173Updated 2 months ago
- "Improving Mathematical Reasoning with Process Supervision" by OPENAI☆108Updated this week
- ☆119Updated 9 months ago
- A benchmark for evaluating learning agents based on just language feedback☆74Updated last month
- ☆102Updated 5 months ago
- The Library for LLM-based multi-agent applications☆80Updated 2 months ago
- augmented LLM with self reflection☆121Updated last year
- An Analytical Evaluation Board of Multi-turn LLM Agents [NeurIPS 2024 Oral]☆313Updated 11 months ago
- ☆114Updated 2 months ago
- Code for ScribeAgent paper☆57Updated 2 months ago
- Resources for our paper: "Agent-R: Training Language Model Agents to Reflect via Iterative Self-Training"☆132Updated last month