raga-ai-hub / raga-llm-hubLinks
Framework for LLM evaluation, guardrails and security
☆112Updated 10 months ago
Alternatives and similar repositories for raga-llm-hub
Users that are interested in raga-llm-hub are comparing it to the libraries listed below
Sorting:
- Recipes and resources for building, deploying, and fine-tuning generative AI with Fireworks.☆117Updated last week
- Fiddler Auditor is a tool to evaluate language models.☆184Updated last year
- Initiative to evaluate and rank the most popular LLMs across common task types based on their propensity to hallucinate.☆111Updated 10 months ago
- ☆73Updated last year
- ☆31Updated 6 months ago
- LangFair is a Python library for conducting use-case level LLM bias and fairness assessments☆219Updated this week
- ☆71Updated 8 months ago
- GenAI Experimentation☆57Updated this week
- A semantic research engine to get relevant papers based on a user query. Application frontend with Chainlit Copilot. Observability with L…☆83Updated last year
- Python SDK for experimenting, testing, evaluating & monitoring LLM-powered applications - Parea AI (YC S23)☆78Updated 5 months ago
- Benchmark various LLM Structured Output frameworks: Instructor, Mirascope, Langchain, LlamaIndex, Fructose, Marvin, Outlines, etc on task…☆173Updated 9 months ago
- Low latency, High Accuracy, Custom Query routers for Humans and Agents. Built by Prithivi Da☆105Updated 3 months ago
- A framework for fine-tuning retrieval-augmented generation (RAG) systems.☆122Updated this week
- Rank LLMs, RAG systems, and prompts using automated head-to-head evaluation☆105Updated 7 months ago
- this project will bootstrap and scaffold the projects for specific semantic search and RAG applications along with regular boiler plate c…☆91Updated 7 months ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated last year
- This repo is the central repo for all the RAG Evaluation reference material and partner workshop☆68Updated 2 months ago
- Sample notebooks and prompts for LLM evaluation☆135Updated last month
- Mistral + Haystack: build RAG pipelines that rock 🤘☆105Updated last year
- LangChain, Llama2-Chat, and zero- and few-shot prompting are used to generate synthetic datasets for IR and RAG system evaluation☆37Updated last year
- Simple AI agents / assistants☆48Updated 9 months ago
- Testing and evaluation framework for voice agents☆128Updated last month
- This repository contains the source code for running llamaindex tutorials from https://howaibuildthis.substack.com/☆41Updated last year
- ☆20Updated last year
- ☆29Updated last year
- A project that enables identification and classification of an intent of a message with dynamic labels☆41Updated 7 months ago
- Multi-Agents using Workflows☆51Updated 6 months ago
- A comprehensive guide to LLM evaluation methods designed to assist in identifying the most suitable evaluation techniques for various use…☆123Updated last week
- ☆89Updated last year
- RAG example using DSPy, Gradio, FastAPI☆83Updated last year