chroma-core / generative-benchmarkingLinks
☆41Updated 2 weeks ago
Alternatives and similar repositories for generative-benchmarking
Users that are interested in generative-benchmarking are comparing it to the libraries listed below
Sorting:
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.☆168Updated last week
- Collection of resources for RL and Reasoning☆26Updated 8 months ago
- Benchmark various LLM Structured Output frameworks: Instructor, Mirascope, Langchain, LlamaIndex, Fructose, Marvin, Outlines, etc on task…☆179Updated last year
- Attribute (or cite) statements generated by LLMs back to in-context information.☆291Updated last year
- Codebase accompanying the Summary of a Haystack paper.☆79Updated last year
- A small library of LLM judges☆294Updated 2 months ago
- ☆146Updated last year
- RAGElo is a set of tools that helps you selecting the best RAG-based LLM agents by using an Elo ranker☆119Updated 3 weeks ago
- Official Repo for CRMArena and CRMArena-Pro☆119Updated 3 months ago
- awesome synthetic (text) datasets☆298Updated 3 months ago
- 🔧 Compare how Agent systems perform on several benchmarks. 📊 🚀☆102Updated 2 months ago
- Python library to use Pleias-RAG models☆63Updated 5 months ago
- ☆78Updated 9 months ago
- 🤗 Benchmark Large Language Models Reliably On Your Data☆404Updated 2 weeks ago
- ☆119Updated last year
- ☆159Updated 10 months ago
- ☆80Updated this week
- Using open source LLMs to build synthetic datasets for direct preference optimization☆66Updated last year
- Banishing LLM Hallucinations Requires Rethinking Generalization☆275Updated last year
- This is the reproduction repository for my 🤗 Hugging Face blog post on synthetic data☆68Updated last year
- Let's build better datasets, together!☆262Updated 9 months ago
- The first dense retrieval model that can be prompted like an LM☆89Updated 5 months ago
- MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents [EMNLP 2024]☆185Updated last month
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆50Updated last year
- Source code of "How to Correctly do Semantic Backpropagation on Language-based Agentic Systems" 🤖☆76Updated 10 months ago
- Generalist and Lightweight Model for Text Classification☆163Updated 4 months ago
- Vision Document Retrieval (ViDoRe): Benchmark. Evaluation code for the ColPali paper.☆245Updated 2 months ago
- Research repository on interfacing LLMs with Weaviate APIs. Inspired by the Berkeley Gorilla LLM.☆135Updated last month
- ☆95Updated 6 months ago
- ☆136Updated last month