shacharKZ / Visualizing-the-Information-Flow-of-GPTLinks
☆11Updated 2 years ago
Alternatives and similar repositories for Visualizing-the-Information-Flow-of-GPT
Users that are interested in Visualizing-the-Information-Flow-of-GPT are comparing it to the libraries listed below
Sorting:
- CausalGym: Benchmarking causal interpretability methods on linguistic tasks☆48Updated 10 months ago
- ☆92Updated last year
- ☆36Updated 3 years ago
- LLM experiments done during SERI MATS - focusing on activation steering / interpreting activation spaces☆98Updated 2 years ago
- Stanford NLP Python library for benchmarking the utility of LLM interpretability methods☆136Updated 3 months ago
- ☆36Updated 2 years ago
- ☆116Updated last year
- Synthetic question-answering dataset to formally analyze the chain-of-thought output of large language models on a reasoning task.☆150Updated last month
- Evaluate interpretability methods on localizing and disentangling concepts in LLMs.☆56Updated last year
- Finding semantically meaningful and accurate prompts.☆48Updated last year
- Investigating the generalization behavior of LM probes trained to predict truth labels: (1) from one annotator to another, and (2) from e…☆28Updated last year
- [ICLR 2023] Code for our paper "Selective Annotation Makes Language Models Better Few-Shot Learners"☆111Updated 2 years ago
- A mechanistic approach for understanding and detecting factual errors of large language models.☆46Updated last year
- Code for preprint: Summarizing Differences between Text Distributions with Natural Language☆43Updated 2 years ago
- A Mechanistic Understanding of Alignment Algorithms: A Case Study on DPO and Toxicity.☆81Updated 7 months ago
- AI Logging for Interpretability and Explainability🔬☆129Updated last year
- ☆48Updated last year
- datasets from the paper "Towards Understanding Sycophancy in Language Models"☆94Updated last year
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆71Updated last year
- ☆56Updated 2 years ago
- ☆233Updated last year
- Skill-It! A Data-Driven Skills Framework for Understanding and Training Language Models☆47Updated last year
- ☆100Updated last year
- [NAACL'25 Oral] Steering Knowledge Selection Behaviours in LLMs via SAE-Based Representation Engineering☆64Updated 10 months ago
- ☆55Updated 2 years ago
- We develop benchmarks and analysis tools to evaluate the causal reasoning abilities of LLMs.☆128Updated last year
- ☆52Updated 6 months ago
- A library for efficient patching and automatic circuit discovery.☆77Updated 2 months ago
- This repository includes code for the paper "Does Localization Inform Editing? Surprising Differences in Where Knowledge Is Stored vs. Ca…☆61Updated 2 years ago
- Repo for: When to Make Exceptions: Exploring Language Models as Accounts of Human Moral Judgment☆38Updated 2 years ago