nrimsky / CAA
Steering Llama 2 with Contrastive Activation Addition
☆97Updated 5 months ago
Related projects ⓘ
Alternatives and complementary repositories for CAA
- LLM experiments done during SERI MATS - focusing on activation steering / interpreting activation spaces☆78Updated last year
- ☆105Updated last month
- Steering vectors for transformer language models in Pytorch / Huggingface☆65Updated last month
- ☆188Updated last month
- A library for efficient patching and automatic circuit discovery.☆31Updated last month
- Sparse probing paper full code.☆51Updated 11 months ago
- ☆107Updated this week
- Using sparse coding to find distributed representations used by neural networks.☆184Updated last year
- ☆170Updated 8 months ago
- Algebraic value editing in pretrained language models☆57Updated last year
- ☆71Updated 3 months ago
- Code and results accompanying the paper "Refusal in Language Models Is Mediated by a Single Direction".☆123Updated last month
- A resource repository for representation engineering in large language models☆54Updated this week
- Function Vectors in Large Language Models (ICLR 2024)☆119Updated last month
- Improving Alignment and Robustness with Circuit Breakers☆154Updated last month
- ☆98Updated 3 months ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆54Updated 2 weeks ago
- ☆44Updated this week
- ☆108Updated last year
- For OpenMOSS Mechanistic Interpretability Team's Sparse Autoencoder (SAE) research.☆48Updated this week
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆157Updated last month
- datasets from the paper "Towards Understanding Sycophancy in Language Models"☆62Updated last year
- Mechanistic Interpretability Visualizations using React☆198Updated 4 months ago
- Röttger et al. (2023): "XSTest: A Test Suite for Identifying Exaggerated Safety Behaviours in Large Language Models"☆63Updated 10 months ago
- ☆76Updated 9 months ago
- Code for my NeurIPS 2024 ATTRIB paper titled "Attribution Patching Outperforms Automated Circuit Discovery"☆26Updated 5 months ago
- Investigating the generalization behavior of LM probes trained to predict truth labels: (1) from one annotator to another, and (2) from e…☆25Updated 5 months ago
- Code for reproducing our paper "Not All Language Model Features Are Linear"☆61Updated last week
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆62Updated 5 months ago
- AI Logging for Interpretability and Explainability🔬☆89Updated 5 months ago