Xmaster6y / lczerolens
π¬ Interpretability for Leela Chess Zero networks.
β12Updated this week
Alternatives and similar repositories for lczerolens:
Users that are interested in lczerolens are comparing it to the libraries listed below
- Mechanistic Interpretability for Transformer Modelsβ50Updated 2 years ago
- Redwood Research's transformer interpretability toolsβ14Updated 2 years ago
- Code for "Evidence of Learned Look-Ahead in a Chess-Playing Neural Network"β20Updated 10 months ago
- β26Updated 11 months ago
- Measuring the situational awareness of language modelsβ34Updated last year
- PyTorch and NNsight implementation of AtP* (Kramar et al 2024, DeepMind)β18Updated 2 months ago
- A library for efficient patching and automatic circuit discovery.β62Updated last month
- Sparse Autoencoder Training Libraryβ47Updated 5 months ago
- β53Updated 6 months ago
- Code and Data Repo for the CoNLL Paper -- Future Lens: Anticipating Subsequent Tokens from a Single Hidden Stateβ18Updated last year
- π§ Starter templates for doing interpretability researchβ67Updated last year
- β30Updated 11 months ago
- β61Updated 2 years ago
- Delphi was the home of a temple to Phoebus Apollo, which famously had the inscription, 'Know Thyself.' This library lets language models β¦β165Updated this week
- This repository contains the code used for the experiments in the paper "Fine-Tuning Enhances Existing Mechanisms: A Case Study on Entityβ¦β25Updated last year
- Arrakis is a library to conduct, track and visualize mechanistic interpretability experiments.β26Updated 3 weeks ago
- Tools for studying developmental interpretability in neural networks.β87Updated 2 months ago
- β124Updated last week
- Code for reproducing our paper "Not All Language Model Features Are Linear"β73Updated 4 months ago
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"β67Updated 9 months ago
- Open source replication of Anthropic's Crosscoders for Model Diffingβ49Updated 5 months ago
- Mechanistic Interpretability Visualizations using Reactβ238Updated 3 months ago
- β63Updated last month
- β121Updated last year
- Multiple datasets for ARC (Abstraction and Reasoning Corpus)β57Updated last week
- Code for my NeurIPS 2024 ATTRIB paper titled "Attribution Patching Outperforms Automated Circuit Discovery"β30Updated 10 months ago
- Keeping language models honest by directly eliciting knowledge encoded in their activations.β197Updated last week
- we got you broβ35Updated 8 months ago
- β43Updated 2 months ago
- datasets from the paper "Towards Understanding Sycophancy in Language Models"β73Updated last year