raga-ai-hub / raga-llm-hub
Framework for LLM evaluation, guardrails and security
☆111Updated 7 months ago
Alternatives and similar repositories for raga-llm-hub:
Users that are interested in raga-llm-hub are comparing it to the libraries listed below
- ☆32Updated 3 months ago
- Sample notebooks and prompts for LLM evaluation☆124Updated this week
- ☆72Updated 6 months ago
- Low latency, High Accuracy, Custom Query routers for Humans and Agents. Built by Prithivi Da☆102Updated 3 weeks ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆49Updated 9 months ago
- Python SDK for experimenting, testing, evaluating & monitoring LLM-powered applications - Parea AI (YC S23)☆76Updated 2 months ago
- Function Calling Benchmark & Testing☆87Updated 9 months ago
- A project that enables identification and classification of an intent of a message with dynamic labels☆38Updated 4 months ago
- ☆1Updated 9 months ago
- This repo is the central repo for all the RAG Evaluation reference material and partner workshop☆64Updated 6 months ago
- this project will bootstrap and scaffold the projects for specific semantic search and RAG applications along with regular boiler plate c…☆89Updated 4 months ago
- A comprehensive guide to LLM evaluation methods designed to assist in identifying the most suitable evaluation techniques for various use…☆114Updated last week
- Fiddler Auditor is a tool to evaluate language models.☆179Updated last year
- GenAI Experimentation☆58Updated this week
- Rank LLMs, RAG systems, and prompts using automated head-to-head evaluation☆103Updated 4 months ago
- ☆29Updated last year
- A curated list of open source repositories for AI Engineers☆111Updated last month
- Mistral + Haystack: build RAG pipelines that rock 🤘☆103Updated last year
- Testing and evaluation framework for voice agents☆110Updated 2 months ago
- Simple AI agents / assistants☆45Updated 6 months ago
- Code for evaluating with Flow-Judge-v0.1 - an open-source, lightweight (3.8B) language model optimized for LLM system evaluations. Crafte…☆66Updated 5 months ago
- ☆72Updated last year
- Testing speed and accuracy of RAG with, and without Cross Encoder Reranker.☆48Updated last year
- Deep Research through Multi-Agents, using GraphRAG☆67Updated 5 months ago
- ☆88Updated last year
- LangFair is a Python library for conducting use-case level LLM bias and fairness assessments☆201Updated this week
- The fastest Trust Layer for AI Agents☆130Updated last month
- This playlab encompasses a multitude of projects crafted through the utilization of Large Language Models, showcasing the versatility and…☆118Updated this week
- Generate Tools and Toolkits from any Python SDK -- no extra code required☆50Updated 5 months ago
- Simple examples using Argilla tools to build AI☆52Updated 5 months ago