autogenai / easy-problems-that-llms-get-wrong
☆45Updated 8 months ago
Alternatives and similar repositories for easy-problems-that-llms-get-wrong
Users that are interested in easy-problems-that-llms-get-wrong are comparing it to the libraries listed below
Sorting:
- ReDel is a toolkit for researchers and developers to build, iterate on, and analyze recursive multi-agent systems. (EMNLP 2024 Demo)☆78Updated 2 months ago
- Simple examples using Argilla tools to build AI☆52Updated 5 months ago
- Training an LLM to use a calculator with multi-turn reinforcement learning, achieving a **62% absolute increase in evaluation accuracy**.☆37Updated last week
- Just a bunch of benchmark logs for different LLMs☆119Updated 9 months ago
- ☆20Updated last year
- LLM-Training-API: Including Embeddings & ReRankers, mergekit, LaserRMT☆27Updated last year
- Lego for GRPO☆28Updated last month
- ☆48Updated 6 months ago
- Train your own SOTA deductive reasoning model☆92Updated 2 months ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated 10 months ago
- ☆66Updated 11 months ago
- The first dense retrieval model that can be prompted like an LM☆72Updated last week
- Source code for our paper: "SelfGoal: Your Language Agents Already Know How to Achieve High-level Goals".☆65Updated 10 months ago
- Matrix (Multi-Agent daTa geneRation Infra and eXperimentation framework) is a versatile engine for multi-agent conversational data genera…☆53Updated this week
- A DSPy-based implementation of the tree of thoughts method (Yao et al., 2023) for generating persuasive arguments☆80Updated 7 months ago
- [ACL 2024] Do Large Language Models Latently Perform Multi-Hop Reasoning?☆65Updated 2 months ago
- Synthetic Data for LLM Fine-Tuning☆115Updated last year
- Mixing Language Models with Self-Verification and Meta-Verification☆104Updated 5 months ago
- Official homepage for "Self-Harmonized Chain of Thought" (NAACL 2025)☆90Updated 3 months ago
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆110Updated 8 months ago
- Code for evaluating with Flow-Judge-v0.1 - an open-source, lightweight (3.8B) language model optimized for LLM system evaluations. Crafte…☆69Updated 6 months ago
- Evaluating LLMs with CommonGen-Lite☆90Updated last year
- Entropy Based Sampling and Parallel CoT Decoding☆17Updated 7 months ago
- The Benefits of a Concise Chain of Thought on Problem Solving in Large Language Models☆22Updated 5 months ago
- ☆37Updated 3 months ago
- Complex Function Calling Benchmark.☆100Updated 3 months ago
- autologic is a Python package that implements the SELF-DISCOVER framework proposed in the paper SELF-DISCOVER: Large Language Models Self…☆57Updated last year
- Function Calling Benchmark & Testing☆87Updated 10 months ago
- A library for benchmarking the Long Term Memory and Continual learning capabilities of LLM based agents. With all the tests and code you…☆70Updated 5 months ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆172Updated 4 months ago