FieteLab / MESH
☆12Updated 2 years ago
Alternatives and similar repositories for MESH
Users that are interested in MESH are comparing it to the libraries listed below
Sorting:
- The Energy Transformer block, in JAX☆57Updated last year
- ☆25Updated 2 years ago
- Official repository for the paper "Exploring the Promise and Limits of Real-Time Recurrent Learning" (ICLR 2024)☆11Updated last year
- Hrrformer: A Neuro-symbolic Self-attention Model (ICML23)☆54Updated last year
- Implementation/simulation of the predictive forward-forward credit assignment algorithm for training neurobiologically-plausible recurren…☆57Updated 2 years ago
- ☆11Updated this week
- A programming language for formal/informal computation.☆41Updated 3 weeks ago
- ☆27Updated 3 years ago
- Engineering the state of RNN language models (Mamba, RWKV, etc.)☆32Updated 11 months ago
- Code for "Meta Learning Backpropagation And Improving It" @ NeurIPS 2021 https://arxiv.org/abs/2012.14905☆31Updated 3 years ago
- ☆14Updated 3 years ago
- repo for code for paper on general theory associative memory models☆18Updated 2 years ago
- ☆32Updated 7 months ago
- Official repository for the paper "Neural Differential Equations for Learning to Program Neural Nets Through Continuous Learning Rules" (…☆22Updated 2 years ago
- ☆31Updated last year
- Universal Neurons in GPT2 Language Models☆29Updated 11 months ago
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆83Updated last year
- This repository includes code to reproduce the tables in "Loss Landscapes are All You Need: Neural Network Generalization Can Be Explaine…☆36Updated 2 years ago
- Parallelizing non-linear sequential models over the sequence length☆51Updated 3 months ago
- Experiments on the impact of depth in transformers and SSMs.☆26Updated 6 months ago
- 🧮 Algebraic Positional Encodings.☆13Updated 4 months ago
- Meta-learning inductive biases in the form of useful conserved quantities.☆37Updated 2 years ago
- An attempt to merge ESBN with Transformers, to endow Transformers with the ability to emergently bind symbols☆15Updated 3 years ago
- ☆39Updated 3 years ago
- HGRN2: Gated Linear RNNs with State Expansion☆54Updated 8 months ago
- Xmixers: A collection of SOTA efficient token/channel mixers☆11Updated 6 months ago
- Deep Networks Grok All the Time and Here is Why☆34Updated 11 months ago
- ☆47Updated last year
- RWKV model implementation☆37Updated last year
- Official repository for the paper "Going Beyond Linear Transformers with Recurrent Fast Weight Programmers" (NeurIPS 2021)☆49Updated 2 years ago