inseq-team / inseqLinks
Interpretability for sequence generation models π π
β433Updated 3 months ago
Alternatives and similar repositories for inseq
Users that are interested in inseq are comparing it to the libraries listed below
Sorting:
- Tools for understanding how transformer predictions are built layer-by-layerβ514Updated 2 weeks ago
- Repository for research in the field of Responsible NLP at Meta.β202Updated 3 months ago
- β331Updated this week
- What's In My Big Data (WIMBD) - a toolkit for analyzing large text datasetsβ223Updated 9 months ago
- Code for T-Few from "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning"β456Updated last year
- Stanford NLP Python library for understanding and improving PyTorch models via interventionsβ793Updated 2 weeks ago
- Materials for EACL2024 tutorial: Transformer-specific Interpretabilityβ60Updated last year
- The official code of LM-Debugger, an interactive tool for inspection and intervention in transformer-based language models.β178Updated 3 years ago
- Utilities for the HuggingFace transformers libraryβ70Updated 2 years ago
- Package to compute Mauve, a similarity score between neural text and human text. Install with `pip install mauve-text`.β295Updated last year
- This repository collects all relevant resources about interpretability in LLMsβ368Updated 9 months ago
- Mechanistic Interpretability Visualizations using Reactβ280Updated 8 months ago
- Aligning AI With Shared Human Values (ICLR 2021)β296Updated 2 years ago
- A Python library that encapsulates various methods for neuron interpretation and analysis in Deep NLP models.β103Updated last year
- A framework for few-shot evaluation of autoregressive language models.β105Updated 2 years ago
- ACL 2022: An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-trained Language Models.β144Updated this week
- PAIR.withgoogle.com and friend's work on interpretability methodsβ198Updated 2 weeks ago
- How do transformer LMs encode relations?β52Updated last year
- Erasing concepts from neural representations with provable guaranteesβ232Updated 6 months ago
- The nnsight package enables interpreting and manipulating the internals of deep learned models.β630Updated this week
- Mistral: A strong, northwesterly wind: Framework for transparent and accessible large-scale language model training, built with Hugging Fβ¦β573Updated last year
- β276Updated last year
- StereoSet: Measuring stereotypical bias in pretrained language modelsβ186Updated 2 years ago
- Organize your experiments into discrete steps that can be cached and reused throughout the lifetime of your research project.β563Updated last year
- A python package for benchmarking interpretability techniques on Transformers.β214Updated 10 months ago
- Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)β462Updated 2 years ago
- Seminar on Large Language Models (COMP790-101 at UNC Chapel Hill, Fall 2022)β311Updated 2 years ago
- Locating and editing factual associations in GPT (NeurIPS 2022)β657Updated last year
- A library for finding knowledge neurons in pretrained transformer models.β158Updated 3 years ago
- PyTorch + HuggingFace code for RetoMaton: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022), including anβ¦β277Updated 2 years ago