safety-research / false-factsLinks
☆30Updated 6 months ago
Alternatives and similar repositories for false-facts
Users that are interested in false-facts are comparing it to the libraries listed below
Sorting:
- Official repo for Learning to Reason for Long-Form Story Generation☆73Updated 8 months ago
- Steering vectors for transformer language models in Pytorch / Huggingface☆137Updated 10 months ago
- ☆92Updated 3 weeks ago
- ☆64Updated this week
- ☆150Updated 4 months ago
- Functional Benchmarks and the Reasoning Gap☆90Updated last year
- ☆104Updated 5 months ago
- ☆105Updated last year
- ☆112Updated 10 months ago
- Measuring the situational awareness of language models☆39Updated last year
- ☆127Updated 2 months ago
- ☆80Updated 3 months ago
- A framework for pitting LLMs against each other in an evolving library of games ⚔☆34Updated 8 months ago
- Archon provides a modular framework for combining different inference-time techniques and LMs with just a JSON config file.☆190Updated 10 months ago
- CausalGym: Benchmarking causal interpretability methods on linguistic tasks☆51Updated last year
- Course Materials for Interpretability of Large Language Models (0368.4264) at Tel Aviv University☆279Updated 3 weeks ago
- Synthetic data generation and benchmark implementation for "Episodic Memories Generation and Evaluation Benchmark for Large Language Mode…☆62Updated 3 months ago
- A toolkit for describing model features and intervening on those features to steer behavior.☆225Updated 3 weeks ago
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆71Updated last year
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆132Updated 3 years ago
- ☆59Updated 11 months ago
- Code and Data Repo for the CoNLL Paper -- Future Lens: Anticipating Subsequent Tokens from a Single Hidden State☆20Updated 2 months ago
- A reading list of relevant papers and projects on foundation model annotation☆28Updated 10 months ago
- Open source interpretability artefacts for R1.☆165Updated 8 months ago
- Understanding how features learned by neural networks evolve throughout training☆41Updated last year
- Sparse Autoencoder Training Library☆56Updated 8 months ago
- An attribution library for LLMs☆46Updated last year
- A toolkit that provides a range of model diffing techniques including a UI to visualize them interactively.☆49Updated this week
- Open Source Replication of Anthropic's Alignment Faking Paper☆52Updated 9 months ago
- datasets from the paper "Towards Understanding Sycophancy in Language Models"☆100Updated 2 years ago