davidbau / baukit
☆158Updated 6 months ago
Related projects: ⓘ
- Function Vectors in Large Language Models (ICLR 2024)☆107Updated last month
- ☆91Updated last month
- ☆174Updated 4 months ago
- Using sparse coding to find distributed representations used by neural networks.☆162Updated 10 months ago
- ☆75Updated this week
- ☆68Updated 10 months ago
- Improving Alignment and Robustness with Circuit Breakers☆124Updated 2 months ago
- AI Logging for Interpretability and Explainability🔬☆74Updated 3 months ago
- LLM experiments done during SERI MATS - focusing on activation steering / interpreting activation spaces☆73Updated 11 months ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆46Updated last month
- Steering Llama 2 with Contrastive Activation Addition☆83Updated 3 months ago
- Algebraic value editing in pretrained language models☆54Updated 10 months ago
- datasets from the paper "Towards Understanding Sycophancy in Language Models"☆59Updated 10 months ago
- Sparse probing paper full code.☆47Updated 9 months ago
- A resource repository for representation engineering in large language models☆35Updated last week
- Inspecting and Editing Knowledge Representations in Language Models☆104Updated last year
- ☆108Updated 3 weeks ago
- Source code of "Task arithmetic in the tangent space: Improved editing of pre-trained models".☆79Updated last year
- The accompanying code for "Transformer Feed-Forward Layers Are Key-Value Memories". Mor Geva, Roei Schuster, Jonathan Berant, and Omer Le…☆80Updated 3 years ago
- ☆63Updated last month
- A library for finding knowledge neurons in pretrained transformer models.☆145Updated 2 years ago
- contrastive decoding☆174Updated last year
- ☆113Updated 2 months ago
- Code for In-context Vectors: Making In Context Learning More Effective and Controllable Through Latent Space Steering☆127Updated 2 months ago
- Code and results accompanying the paper "Refusal in Language Models Is Mediated by a Single Direction".☆75Updated 3 weeks ago
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).☆131Updated last month
- Röttger et al. (2023): "XSTest: A Test Suite for Identifying Exaggerated Safety Behaviours in Large Language Models"☆54Updated 8 months ago
- ☆68Updated 7 months ago
- For OpenMOSS Mechanistic Interpretability Team's Sparse Autoencoder (SAE) research.☆30Updated last week
- ☆135Updated 4 months ago