inseq-team / inseqLinks
Interpretability for sequence generation models π π
β439Updated 2 weeks ago
Alternatives and similar repositories for inseq
Users that are interested in inseq are comparing it to the libraries listed below
Sorting:
- Stanford NLP Python library for understanding and improving PyTorch models via interventionsβ813Updated last month
- Tools for understanding how transformer predictions are built layer-by-layerβ527Updated last month
- Repository for research in the field of Responsible NLP at Meta.β202Updated 4 months ago
- Utilities for the HuggingFace transformers libraryβ72Updated 2 years ago
- What's In My Big Data (WIMBD) - a toolkit for analyzing large text datasetsβ224Updated 10 months ago
- β361Updated last week
- Materials for EACL2024 tutorial: Transformer-specific Interpretabilityβ60Updated last year
- The official code of LM-Debugger, an interactive tool for inspection and intervention in transformer-based language models.β179Updated 3 years ago
- Erasing concepts from neural representations with provable guaranteesβ237Updated 8 months ago
- PAIR.withgoogle.com and friend's work on interpretability methodsβ203Updated last week
- Aligning AI With Shared Human Values (ICLR 2021)β299Updated 2 years ago
- This repository collects all relevant resources about interpretability in LLMsβ373Updated 11 months ago
- Code for T-Few from "Few-Shot Parameter-Efficient Fine-Tuning is Better and Cheaper than In-Context Learning"β457Updated 2 years ago
- A Python library that encapsulates various methods for neuron interpretation and analysis in Deep NLP models.β105Updated last year
- Package to compute Mauve, a similarity score between neural text and human text. Install with `pip install mauve-text`.β298Updated last year
- How do transformer LMs encode relations?β53Updated last year
- Locating and editing factual associations in GPT (NeurIPS 2022)β671Updated last year
- A framework for few-shot evaluation of autoregressive language models.β105Updated 2 years ago
- Mistral: A strong, northwesterly wind: Framework for transparent and accessible large-scale language model training, built with Hugging Fβ¦β575Updated last year
- Mechanistic Interpretability Visualizations using Reactβ289Updated 9 months ago
- β276Updated last year
- β242Updated last year
- Organize your experiments into discrete steps that can be cached and reused throughout the lifetime of your research project.β563Updated last year
- β81Updated 7 months ago
- Datasets collection and preprocessings framework for NLP extreme multitask learningβ186Updated 2 months ago
- StereoSet: Measuring stereotypical bias in pretrained language modelsβ191Updated 2 years ago
- Seminar on Large Language Models (COMP790-101 at UNC Chapel Hill, Fall 2022)β311Updated 2 years ago
- Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)β462Updated 2 years ago
- PyTorch + HuggingFace code for RetoMaton: "Neuro-Symbolic Language Modeling with Automaton-augmented Retrieval" (ICML 2022), including anβ¦β280Updated 2 years ago
- ACL 2022: An Empirical Survey of the Effectiveness of Debiasing Techniques for Pre-trained Language Models.β147Updated last month