stephenleo / llm-structured-output-benchmarksLinks
Benchmark various LLM Structured Output frameworks: Instructor, Mirascope, Langchain, LlamaIndex, Fructose, Marvin, Outlines, etc on tasks like multi-label classification, named entity recognition, synthetic data generation, etc.
☆178Updated 11 months ago
Alternatives and similar repositories for llm-structured-output-benchmarks
Users that are interested in llm-structured-output-benchmarks are comparing it to the libraries listed below
Sorting:
- ARAGOG- Advanced RAG Output Grading. Exploring and comparing various Retrieval-Augmented Generation (RAG) techniques on AI research paper…☆109Updated last year
- ☆235Updated 3 months ago
- RAGElo is a set of tools that helps you selecting the best RAG-based LLM agents by using an Elo ranker☆115Updated last week
- In-Context Learning for eXtreme Multi-Label Classification (XMC) using only a handful of examples.☆437Updated last year
- Repository to demonstrate Chain of Table reasoning with multiple tables powered by LangGraph☆147Updated last year
- ☆145Updated last year
- Recipes for learning, fine-tuning, and adapting ColPali to your multimodal RAG use cases. 👨🏻🍳☆331Updated 3 months ago
- Mistral + Haystack: build RAG pipelines that rock 🤘☆105Updated last year
- Vision Document Retrieval (ViDoRe): Benchmark. Evaluation code for the ColPali paper.☆236Updated last month
- Building a chatbot powered with a RAG pipeline to read,summarize and quote the most relevant papers related to the user query.☆168Updated last year
- Generalist and Lightweight Model for Text Classification☆157Updated 3 months ago
- A Lightweight Library for AI Observability☆251Updated 6 months ago
- Domain Adapted Language Modeling Toolkit - E2E RAG☆328Updated 10 months ago
- Function Calling Benchmark & Testing☆89Updated last year
- Low latency, High Accuracy, Custom Query routers for Humans and Agents. Built by Prithivi Da☆116Updated 5 months ago
- Simple UI for debugging correlations of text embeddings☆291Updated 3 months ago
- Code for evaluating with Flow-Judge-v0.1 - an open-source, lightweight (3.8B) language model optimized for LLM system evaluations. Crafte…☆78Updated 10 months ago
- This repository implements the chain of verification paper by Meta AI☆176Updated last year
- Experimental Code for StructuredRAG: JSON Response Formatting with Large Language Models☆111Updated 5 months ago
- FastAPI wrapper around DSPy☆267Updated last year
- LLM-driven automated knowledge graph construction from text using DSPy and Neo4j.☆190Updated last year
- ☆210Updated 2 months ago
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆114Updated last month
- Synthetic Data for LLM Fine-Tuning☆120Updated last year
- Model, Code & Data for the EMNLP'23 paper "Making Large Language Models Better Data Creators"☆135Updated last year
- A small library of LLM judges☆282Updated last month
- Attribute (or cite) statements generated by LLMs back to in-context information.☆280Updated 11 months ago
- Banishing LLM Hallucinations Requires Rethinking Generalization☆276Updated last year
- 📚 Datasets and models for instruction-tuning☆238Updated last year
- DSPY on action with OpenSource LLMs.☆90Updated last year