catherinesyeh / attention-vizLinks
Visualizing query-key interactions in language + vision transformers (VIS 2023)
☆158Updated last year
Alternatives and similar repositories for attention-viz
Users that are interested in attention-viz are comparing it to the libraries listed below
Sorting:
- Extracting spatial and temporal world models from LLMs☆257Updated 2 years ago
- ☆112Updated 11 months ago
- ☆152Updated 4 months ago
- Emergent world representations: Exploring a sequence model trained on a synthetic task☆199Updated 2 years ago
- ☆105Updated last year
- Website for hosting the Open Foundation Models Cheat Sheet.☆269Updated 8 months ago
- Implementation of CALM from the paper "LLM Augmented LLMs: Expanding Capabilities through Composition", out of Google Deepmind☆179Updated last year
- A mechanistic approach for understanding and detecting factual errors of large language models.☆49Updated last year
- LLM-Merging: Building LLMs Efficiently through Merging☆209Updated last year
- RuLES: a benchmark for evaluating rule-following in language models☆247Updated 11 months ago
- Notebooks accompanying Anthropic's "Toy Models of Superposition" paper☆133Updated 3 years ago
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆198Updated last year