amazon-science / llm-interpret
Code for the ACL 2023 paper: "Rethinking the Role of Scale for In-Context Learning: An Interpretability-based Case Study at 66 Billion Scale"
☆28Updated last year
Related projects ⓘ
Alternatives and complementary repositories for llm-interpret
- Fast and Robust Early-Exiting Framework for Autoregressive Language Models with Synchronized Parallel Decoding (EMNLP 2023 Long)☆53Updated last month
- ☆31Updated last year
- ☆71Updated 6 months ago
- ☆38Updated 7 months ago
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆41Updated 10 months ago
- ☆46Updated this week
- Data and code for the preprint "In-Context Learning with Long-Context Models: An In-Depth Exploration"☆31Updated 3 months ago
- The accompanying code for "Transformer Feed-Forward Layers Are Key-Value Memories". Mor Geva, Roei Schuster, Jonathan Berant, and Omer Le…☆85Updated 3 years ago
- CausalGym: Benchmarking causal interpretability methods on linguistic tasks☆30Updated 8 months ago
- Few-shot Learning with Auxiliary Data☆26Updated 11 months ago
- ☆122Updated 10 months ago
- ☆32Updated last year
- SILO Language Models code repository☆80Updated 8 months ago
- This repository contains the code used for the experiments in the paper "Fine-Tuning Enhances Existing Mechanisms: A Case Study on Entity…☆19Updated 8 months ago
- Language models scale reliably with over-training and on downstream tasks☆94Updated 7 months ago
- Repository for "Propagating Knowledge Updates to LMs Through Distillation" (NeurIPS 2023).☆24Updated 2 months ago
- ☆58Updated last year
- Code for the ICLR 2024 paper "How to catch an AI liar: Lie detection in black-box LLMs by asking unrelated questions"☆62Updated 5 months ago
- ☆36Updated 3 months ago
- Codebase for Instruction Following without Instruction Tuning☆32Updated last month
- Long Context Extension and Generalization in LLMs☆39Updated 2 months ago
- ☆47Updated 9 months ago
- ☆44Updated 2 months ago
- ☆34Updated 3 months ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models"☆56Updated last month
- Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"☆56Updated last year
- Simple and efficient pytorch-native transformer training and inference (batched)☆61Updated 7 months ago
- Code release for Dataless Knowledge Fusion by Merging Weights of Language Models (https://openreview.net/forum?id=FCnohuR6AnM)☆85Updated last year
- FollowIR: Evaluating and Teaching Information Retrieval Models to Follow Instructions☆40Updated 4 months ago
- Adding new tasks to T0 without catastrophic forgetting☆30Updated 2 years ago