inseq-team / inseq
Interpretability for sequence generation models π π
β409Updated 4 months ago
Alternatives and similar repositories for inseq:
Users that are interested in inseq are comparing it to the libraries listed below
- Tools for understanding how transformer predictions are built layer-by-layerβ480Updated 9 months ago
- Datasets collection and preprocessings framework for NLP extreme multitask learningβ176Updated 2 months ago
- What's In My Big Data (WIMBD) - a toolkit for analyzing large text datasetsβ213Updated 4 months ago
- Repository for research in the field of Responsible NLP at Meta.β196Updated 3 months ago
- This repository collects all relevant resources about interpretability in LLMsβ327Updated 4 months ago
- Erasing concepts from neural representations with provable guaranteesβ226Updated last month
- The official code of LM-Debugger, an interactive tool for inspection and intervention in transformer-based language models.β177Updated 2 years ago
- A python package for benchmarking interpretability techniques on Transformers.β213Updated 5 months ago
- Stanford NLP Python library for understanding and improving PyTorch models via interventionsβ721Updated last month
- Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)β463Updated 2 years ago
- Code for T-Few from "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning"β447Updated last year
- Mechanistic Interpretability Visualizations using Reactβ235Updated 3 months ago
- Tools for checking ACL paper submissionsβ675Updated 5 months ago
- β263Updated last year
- Training Sparse Autoencoders on Language Modelsβ669Updated this week
- Create feature-centric and prompt-centric visualizations for sparse autoencoders (like those from Anthropic's published research).β189Updated 3 months ago
- This is the repository of HaluEval, a large-scale hallucination evaluation benchmark for Large Language Models.β451Updated last year
- β241Updated this week
- PyTorch + HuggingFace code for RetoMaton: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022), including anβ¦β270Updated 2 years ago
- A library for finding knowledge neurons in pretrained transformer models.β155Updated 3 years ago
- Mistral: A strong, northwesterly wind: Framework for transparent and accessible large-scale language model training, built with Hugging Fβ¦β569Updated last year
- The nnsight package enables interpreting and manipulating the internals of deep learned models.β522Updated this week
- OpenICL is an open-source framework to facilitate research, development, and prototyping of in-context learning.β552Updated last year
- β211Updated 5 months ago
- β195Updated last year
- SelfCheckGPT: Zero-Resource Black-Box Hallucination Detection for Generative Large Language Modelsβ501Updated 8 months ago
- Utilities for the HuggingFace transformers libraryβ66Updated 2 years ago
- PAIR.withgoogle.com and friend's work on interpretability methodsβ170Updated this week
- Sparse Autoencoder for Mechanistic Interpretabilityβ233Updated 8 months ago
- Powerful unsupervised domain adaptation method for dense retrieval. Requires only unlabeled corpus and yields massive improvement: "GPL: β¦β329Updated last year