aiverify-foundation / LLM-Evals-Catalogue
This repository stems from our paper, “Cataloguing LLM Evaluations”, and serves as a living, collaborative catalogue of LLM evaluation frameworks, benchmarks and papers.
☆17Updated last year
Alternatives and similar repositories for LLM-Evals-Catalogue
Users that are interested in LLM-Evals-Catalogue are comparing it to the libraries listed below
Sorting:
- ARAGOG- Advanced RAG Output Grading. Exploring and comparing various Retrieval-Augmented Generation (RAG) techniques on AI research paper…☆103Updated last year
- A comprehensive guide to LLM evaluation methods designed to assist in identifying the most suitable evaluation techniques for various use…☆116Updated last week
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated 10 months ago
- Sample notebooks and prompts for LLM evaluation☆126Updated last week
- ☆72Updated 6 months ago
- 🔧 Compare how Agent systems perform on several benchmarks. 📊🚀☆96Updated 6 months ago
- A blueprint for AI development, focusing on applied examples of RAG, information extraction, analysis and fine-tuning in the age of LLMs …☆53Updated 3 months ago
- Source code for our paper: "SelfGoal: Your Language Agents Already Know How to Achieve High-level Goals".☆65Updated 10 months ago
- This repository contains a pipeline for fine-tuning Large Language Models (LLMs) for Text-to-SQL conversion using General Reward Proximal…☆21Updated last month
- Notebooks and articles related to LLMs☆25Updated last year
- Simple examples using Argilla tools to build AI☆52Updated 5 months ago
- Cookbooks showcasing various applications of Cleanlab☆15Updated last month
- ☆143Updated 9 months ago
- ☆29Updated last year
- A semantic research engine to get relevant papers based on a user query. Application frontend with Chainlit Copilot. Observability with L…☆82Updated last year
- Lean implementation of various multi-agent LLM methods, including Iteration of Thought (IoT)☆110Updated 3 months ago
- LangChain, Llama2-Chat, and zero- and few-shot prompting are used to generate synthetic datasets for IR and RAG system evaluation☆36Updated last year
- All code related to medium articles☆17Updated this week
- ☆65Updated 2 months ago
- Repository to demonstrate Chain of Table reasoning with multiple tables powered by LangGraph☆144Updated last year
- Research repository on interfacing LLMs with Weaviate APIs. Inspired by the Berkeley Gorilla LLM.☆124Updated 3 weeks ago
- ☆36Updated 10 months ago
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆109Updated 8 months ago
- Fiddler Auditor is a tool to evaluate language models.☆179Updated last year
- Truth Forest: Toward Multi-Scale Truthfulness in Large Language Models through Intervention without Tuning☆46Updated last year
- This repo is the central repo for all the RAG Evaluation reference material and partner workshop☆64Updated 3 weeks ago
- GenAI Experimentation☆58Updated 3 weeks ago
- Model, Code & Data for the EMNLP'23 paper "Making Large Language Models Better Data Creators"☆132Updated last year
- LangChain chat model abstractions for dynamic failover, load balancing, chaos engineering, and more!☆81Updated last year
- Mistral + Haystack: build RAG pipelines that rock 🤘☆103Updated last year