IBM / ICX360Links
In-Context Explainability 360 toolkit
☆51Updated this week
Alternatives and similar repositories for ICX360
Users that are interested in ICX360 are comparing it to the libraries listed below
Sorting:
- Mellea is a library for writing generative programs.☆260Updated last week
- The Agent Lifecycle Toolkit (ALTK) is a library of components to help agent builders improve their agent with minimal integration effort …☆90Updated this week
- The AI Steerability 360 toolkit is an extensible library for general purpose steering of LLMs.☆54Updated last month
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆212Updated this week
- EvalAssist is an open-source project that simplifies using large language models as evaluators (LLM-as-a-Judge) of the output of other la…☆92Updated 3 weeks ago
- Code for "Utility Engineering: Analyzing and Controlling Emergent Value Systems in AIs"☆85Updated 9 months ago
- AI Atlas Nexus: tooling to bring together resources related to governance of foundation models.☆111Updated this week
- Collection of evals for Inspect AI☆313Updated this week
- AssetOpsBench - Industry 4.0☆571Updated last week
- ☆258Updated 8 months ago
- An open-source compliance-centered evaluation framework for Generative AI models☆176Updated this week
- Probabilistic programming with large language models☆154Updated last month
- A toolkit for describing model features and intervening on those features to steer behavior.☆221Updated last week
- Contains random samples referenced in the paper "Sleeper Agents: Training Robustly Deceptive LLMs that Persist Through Safety Training".☆122Updated last year
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.☆173Updated 2 weeks ago
- Steering vectors for transformer language models in Pytorch / Huggingface☆132Updated 9 months ago
- Persona Vectors: Monitoring and Controlling Character Traits in Language Models☆307Updated 4 months ago
- ☆112Updated 10 months ago
- The Granite Guardian models are designed to detect risks in prompts and responses.☆123Updated 2 months ago
- Sparse Autoencoder for Mechanistic Interpretability☆286Updated last year
- ☆88Updated last week
- ControlArena is a collection of settings, model organisms and protocols - for running control experiments.☆135Updated last week
- Python framework which enables you to transform how a user calls or infers an IBM Granite model and how the output from the model is retu…☆51Updated this week
- ☆49Updated last year
- CUGA is an open-source generalist agent for the enterprise, supporting complex task execution on web and APIs, OpenAPI/MCP integrations, …☆388Updated this week
- 🪄 Interpreto is an interpretability toolbox for LLMs☆84Updated last week
- ☆44Updated last year
- ☆401Updated last week
- open source interpretability platform 🧠☆562Updated this week