Fiddler Auditor is a tool to evaluate language models.
β189Mar 11, 2024Updated 2 years ago
Alternatives and similar repositories for fiddler-auditor
Users that are interested in fiddler-auditor are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Research notes and extra resources for all the work at explodinggradients.comβ25Mar 11, 2025Updated last year
- π LangKit: An open-source toolkit for monitoring Large Language Models (LLMs). π Extracts signals from prompts & responses, ensuring saβ¦β977Nov 22, 2024Updated last year
- Scan your AI/ML models for problems before you put them into production.β11Mar 31, 2025Updated 11 months ago
- LLM evaluation.β16Nov 7, 2023Updated 2 years ago
- Python package for dimension reduction of high-dimensional biological data.β16May 23, 2018Updated 7 years ago
- A tool for evaluating LLMsβ428Mar 15, 2026Updated last week
- Just a bunch of benchmark logs for different LLMsβ119Jul 28, 2024Updated last year
- Jump ReLUβ11Apr 8, 2019Updated 6 years ago
- Evaluating LLMs with CommonGen-Liteβ95Mar 21, 2024Updated 2 years ago
- π Unstructured Data Connectors for Haystack 2.0β17Sep 21, 2023Updated 2 years ago
- A complete guide to evaluate LLMs and RAGs. Both theory and code based approaches covered.β29Nov 16, 2023Updated 2 years ago
- Sample project to get started with dbt-power-user vscode extension using dev-containerβ11Apr 5, 2024Updated last year
- All Data Toolsβ10Feb 28, 2023Updated 3 years ago
- Custom Frappe App to integrate AI Agents as DocTypes using CrewAI & Frappe Framework.β42Apr 30, 2024Updated last year
- Deepchecks: Tests for Continuous Validation of ML Models & Data. Deepchecks is a holistic open-source solution for all of your AI & ML vaβ¦β3,994Dec 28, 2025Updated 2 months ago
- First steps to interact with MLflow (mlflow.org)β10Oct 18, 2018Updated 7 years ago
- β12Jun 10, 2024Updated last year
- Sample notebooks and prompts for LLM evaluationβ161Nov 2, 2025Updated 4 months ago
- π’ Open-Source Evaluation & Testing library for LLM Agentsβ5,184Updated this week
- β12Jun 17, 2024Updated last year
- The Security Toolkit for LLM Interactionsβ2,699Dec 15, 2025Updated 3 months ago
- AI Observability & Evaluationβ8,904Updated this week
- Scalable Meta-Evaluation of LLMs as Evaluatorsβ43Feb 15, 2024Updated 2 years ago
- Python SDK for running evaluations on LLM generated responsesβ298Jun 6, 2025Updated 9 months ago
- Sample solution to automate tedious regulatory compliance processes using multi-agent systemsβ24Apr 15, 2025Updated 11 months ago
- sktime workshops & tutorialsβ14Jul 14, 2021Updated 4 years ago
- Template for model cardsβ22Jan 24, 2023Updated 3 years ago
- Adding guardrails to large language models.β6,553Updated this week
- LLM Prompt Injection Detectorβ1,445Aug 7, 2024Updated last year
- NeMo Guardrails is an open-source toolkit for easily adding programmable guardrails to LLM-based conversational systems.β5,819Updated this week
- A content inspecting SMTP proxyβ17Jun 9, 2014Updated 11 years ago
- Awesome material(papers, tools, etc.) about testing machine learning system, including deep learning system.β47Oct 12, 2021Updated 4 years ago
- Building effective agents patterns implemented with Dapr Agentsβ11May 21, 2025Updated 10 months ago
- Solving tic-tac-toe using deep reinforcement learningβ12Jan 26, 2019Updated 7 years ago
- Code for the papers "Induction of Subgoal Automata for Reinforcement Learning" (AAAI-20) and "Induction and Exploitation of Subgoal Automβ¦β14Aug 15, 2023Updated 2 years ago
- the LLM vulnerability scannerβ7,312Updated this week
- Metrics to evaluate the quality of responses of your Retrieval Augmented Generation (RAG) applications.β325Jul 10, 2025Updated 8 months ago
- Homepage for the Data Interaction Group at CMUβ13Updated this week
- Evidently is ββan open-source ML and LLM observability framework. Evaluate, test, and monitor any AI-powered system or data pipeline. Froβ¦β7,308Mar 10, 2026Updated last week