abietti / transformer-birthLinks
☆19Updated last year
Alternatives and similar repositories for transformer-birth
Users that are interested in transformer-birth are comparing it to the libraries listed below
Sorting:
- ☆33Updated last year
- Curse-of-memory phenomenon of RNNs in sequence modelling☆18Updated 5 months ago
- ☆108Updated 8 months ago
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆68Updated last year
- ☆53Updated last year
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆80Updated 2 years ago
- ☆36Updated 3 years ago
- Code for the "Overcoming Sparsity Artifacts in Crosscoders to Interpret Chat-Tuning" paper.☆14Updated 2 weeks ago
- Official repository for our paper, Transformers Learn Higher-Order Optimization Methods for In-Context Learning: A Study with Linear Mode…☆19Updated 11 months ago
- ☆49Updated 9 months ago
- ☆71Updated 10 months ago
- ☆103Updated last year
- Experiments on the impact of depth in transformers and SSMs.☆36Updated last week
- Deep Learning & Information Bottleneck☆61Updated 2 years ago
- ☆31Updated last year
- Implementations of various linear RNN layers using pytorch and triton☆54Updated 2 years ago
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆59Updated 2 years ago
- ☆240Updated last year
- Experiments and code to generate the GINC small-scale in-context learning dataset from "An Explanation for In-context Learning as Implici…☆106Updated last year
- Universal Neurons in GPT2 Language Models☆30Updated last year
- Stick-breaking attention☆61Updated 3 months ago
- Evaluate interpretability methods on localizing and disentangling concepts in LLMs.☆56Updated last year
- Official repo for the paper "Weight-based Decomposition: A Case for Bilinear MLPs"☆23Updated 2 months ago
- A Kernel-Based View of Language Model Fine-Tuning https://arxiv.org/abs/2210.05643☆78Updated 2 years ago
- ☆22Updated 6 months ago
- Parallelizing non-linear sequential models over the sequence length☆54Updated 4 months ago
- Official code for the paper "Attention as a Hypernetwork"☆44Updated last year
- ☆108Updated 2 years ago
- Omnigrok: Grokking Beyond Algorithmic Data☆62Updated 2 years ago
- ☆33Updated last year