IBM / ICX360Links
In-Context Explainability 360 toolkit
☆51Updated this week
Alternatives and similar repositories for ICX360
Users that are interested in ICX360 are comparing it to the libraries listed below
Sorting:
- EvalAssist is an open-source project that simplifies using large language models as evaluators (LLM-as-a-Judge) of the output of other la…☆92Updated last month
- Mellea is a library for writing generative programs.☆271Updated this week
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆212Updated this week
- Governance of the Commons Simulation (GovSim)☆64Updated 11 months ago
- Code for "Utility Engineering: Analyzing and Controlling Emergent Value Systems in AIs"☆86Updated 10 months ago
- The AI Steerability 360 toolkit is an extensible library for general purpose steering of LLMs.☆55Updated 2 months ago
- ☆261Updated 9 months ago
- ☆43Updated last year
- open source interpretability platform 🧠☆621Updated this week
- The Agent Lifecycle Toolkit (ALTK) is a library of components to help agent builders improve their agent with minimal integration effort …☆103Updated this week
- Collection of evals for Inspect AI☆332Updated this week
- Repository for the "Chain-of-Thought Reasoning In The Wild Is Not Always Faithful" paper☆31Updated last month
- ☆191Updated 6 months ago
- Steering vectors for transformer language models in Pytorch / Huggingface☆137Updated 10 months ago
- AI Atlas Nexus: tooling to bring together resources related to governance of foundation models.☆112Updated 2 weeks ago
- The Granite Guardian models are designed to detect risks in prompts and responses.☆126Updated 3 months ago
- A toolkit for describing model features and intervening on those features to steer behavior.☆225Updated last month
- ☆92Updated 3 weeks ago
- ControlArena is a collection of settings, model organisms and protocols - for running control experiments.☆145Updated 3 weeks ago
- This repository collects all relevant resources about interpretability in LLMs☆389Updated last year
- AI Verify☆39Updated this week
- Persona Vectors: Monitoring and Controlling Character Traits in Language Models☆322Updated 5 months ago
- ☆414Updated this week
- Sparse Autoencoder for Mechanistic Interpretability☆285Updated last year
- ⚓️ Repository for the "Thought Anchors: Which LLM Reasoning Steps Matter?" paper.☆104Updated 2 months ago
- Edu-ConvoKit: An Open-Source Framework for Education Conversation Data☆105Updated 8 months ago
- Attribute (or cite) statements generated by LLMs back to in-context information.☆313Updated last year
- An open-source compliance-centered evaluation framework for Generative AI models☆178Updated 2 weeks ago
- Experimental library integrating LLM capabilities to support causal analyses☆278Updated this week
- Open source interpretability artefacts for R1.☆165Updated 8 months ago