IBM / ICX360Links
In-Context Explainability 360 toolkit
☆48Updated 2 weeks ago
Alternatives and similar repositories for ICX360
Users that are interested in ICX360 are comparing it to the libraries listed below
Sorting:
- The AI Steerability 360 toolkit is an extensible library for general purpose steering of LLMs.☆49Updated last month
- EvalAssist is an open-source project that simplifies using large language models as evaluators (LLM-as-a-Judge) of the output of other la…☆92Updated last week
- AssetOpsBench - Industry 4.0☆435Updated last week
- Mellea is a library for writing generative programs.☆234Updated last week
- The Agent Lifecycle Toolkit (ALTK) is a library of components to help agent builders improve their agent with minimal integration effort …☆78Updated this week
- Risk Atlas Nexus: tooling to bring together resources related to governance of foundation models.☆110Updated 2 weeks ago
- CUGA is an open-source generalist agent for the enterprise, supporting complex task execution on web and APIs, OpenAPI/MCP integrations, …☆207Updated last week
- An open-source compliance-centered evaluation framework for Generative AI models☆172Updated 2 weeks ago
- The Granite Guardian models are designed to detect risks in prompts and responses.☆121Updated last month
- Benchmarks for the Evaluation of LLM Supervision☆32Updated last month
- Prompt Declaration Language (PDL) is a declarative prompt programming language.☆266Updated this week
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆212Updated last week
- Python framework which enables you to transform how a user calls or infers an IBM Granite model and how the output from the model is retu…☆51Updated this week
- Collection of evals for Inspect AI☆290Updated this week
- Experimental library integrating LLM capabilities to support causal analyses☆259Updated last month
- TalkToModel gives anyone with the powers of XAI through natural language conversations 💬!☆125Updated 2 years ago
- This repository stems from our paper, “Cataloguing LLM Evaluations”, and serves as a living, collaborative catalogue of LLM evaluation fr…☆18Updated 2 years ago
- LangFair is a Python library for conducting use-case level LLM bias and fairness assessments☆242Updated last week
- This is an open-source tool to assess and improve the trustworthiness of AI systems.☆98Updated this week
- Guardrails for secure and robust agent development☆366Updated 4 months ago
- PyTorch package to train and audit ML models for Individual Fairness☆66Updated 2 months ago
- Run safety benchmarks against AI models and view detailed reports showing how well they performed.☆111Updated this week
- ☆49Updated last year
- A Comprehensive Assessment of Trustworthiness in GPT Models☆308Updated last year
- AI Verify☆37Updated 3 weeks ago
- A curated list of awesome academic research, books, code of ethics, courses, databases, data sets, frameworks, institutes, maturity mode…☆95Updated this week
- A curated list of awesome synthetic data tools (open source and commercial).☆222Updated last year
- Granite Snack Cookbook -- easily consumable recipes (python notebooks) that showcase the capabilities of the Granite models☆317Updated this week
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".☆122Updated last year
- ☆257Updated 8 months ago