MadryLab / context-citeLinks
Attribute (or cite) statements generated by LLMs back to in-context information.
☆319Updated last year
Alternatives and similar repositories for context-cite
Users that are interested in context-cite are comparing it to the libraries listed below
Sorting:
- awesome synthetic (text) datasets☆321Updated 3 weeks ago
- Code accompanying "How I learned to start worrying about prompt formatting".☆113Updated 7 months ago
- Dense X Retrieval: What Retrieval Granularity Should We Use?☆168Updated 2 years ago
- MiniCheck: Efficient Fact-Checking of LLMs on Grounding Documents [EMNLP 2024]☆194Updated 5 months ago
- Code repo for "Agent Instructs Large Language Models to be General Zero-Shot Reasoners"☆120Updated 3 months ago
- ☆120Updated last year
- Official repository for "Scaling Retrieval-Based Langauge Models with a Trillion-Token Datastore".☆224Updated last month
- A small library of LLM judges☆319Updated 6 months ago
- Vision Document Retrieval (ViDoRe): Benchmark. Evaluation code for the ColPali paper.☆259Updated 2 weeks ago
- The first dense retrieval model that can be prompted like an LM☆90Updated 8 months ago
- ☆147Updated last year
- Comprehensive benchmark for RAG☆260Updated 7 months ago
- Steer LLM outputs towards a certain topic/subject and enhance response capabilities using activation engineering by adding steering vecto…☆268Updated 3 weeks ago
- Benchmark various LLM Structured Output frameworks: Instructor, Mirascope, Langchain, LlamaIndex, Fructose, Marvin, Outlines, etc on task…☆184Updated last year
- Evaluating LLMs with fewer examples☆169Updated last year
- Official Implementation of "Multi-Head RAG: Solving Multi-Aspect Problems with LLMs"☆237Updated 4 months ago
- Codebase accompanying the Summary of a Haystack paper.☆80Updated last year
- RAGElo is a set of tools that helps you selecting the best RAG-based LLM agents by using an Elo ranker☆126Updated 3 months ago
- ☆130Updated last year
- Official repository for paper "ReasonIR Training Retrievers for Reasoning Tasks".☆217Updated 7 months ago
- Complex Function Calling Benchmark.☆163Updated last year
- In-Context Learning for eXtreme Multi-Label Classification (XMC) using only a handful of examples.☆446Updated last year
- The official evaluation suite and dynamic data release for MixEval.☆255Updated last year
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"☆223Updated last year
- 🤗 Benchmark Large Language Models Reliably On Your Data☆426Updated last month
- Official repo for "LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs".☆241Updated last year
- A comprehensive guide to LLM evaluation methods designed to assist in identifying the most suitable evaluation techniques for various use…☆176Updated 2 weeks ago
- 🔧 Compare how Agent systems perform on several benchmarks. 📊🚀☆103Updated 6 months ago
- ☆161Updated last year
- Official repo for "Make Your LLM Fully Utilize the Context"☆263Updated last year