Ads97 / ForwardForward
Explorations with Geoffrey Hinton's Forward Forward algoithm
☆33Updated 10 months ago
Related projects ⓘ
Alternatives and complementary repositories for ForwardForward
- Reimplementation of Geoffrey Hinton's Forward-Forward Algorithm☆132Updated last year
- Implementation/simulation of the predictive forward-forward credit assignment algorithm for training neurobiologically-plausible recurren…☆55Updated last year
- ☆58Updated 2 years ago
- ☆53Updated 3 months ago
- Code for "Meta Learning Backpropagation And Improving It" @ NeurIPS 2021 https://arxiv.org/abs/2012.14905☆31Updated 2 years ago
- The Energy Transformer block, in JAX☆53Updated 11 months ago
- ☆48Updated 9 months ago
- Parallelizing non-linear sequential models over the sequence length☆45Updated 3 weeks ago
- HGRN2: Gated Linear RNNs with State Expansion☆49Updated 3 months ago
- This repository includes code to reproduce the tables in "Loss Landscapes are All You Need: Neural Network Generalization Can Be Explaine…☆34Updated last year
- Hrrformer: A Neuro-symbolic Self-attention Model (ICML23)☆47Updated last year
- Implementation of "Gradients without backpropagation" paper (https://arxiv.org/abs/2202.08587) using functorch☆96Updated last year
- ☆45Updated 9 months ago
- ☆29Updated 2 months ago
- ☆36Updated 2 years ago
- ☆31Updated 10 months ago
- [ICLR 2023] "Sparsity May Cry: Let Us Fail (Current) Sparse Neural Networks Together!" Shiwei Liu, Tianlong Chen, Zhenyu Zhang, Xuxi Chen…☆27Updated last year
- ☆28Updated 7 months ago
- Official implementation of the transformer (TF) architecture suggested in a paper entitled "Looped Transformers as Programmable Computers…☆22Updated last year
- Official repository for the paper "Going Beyond Linear Transformers with Recurrent Fast Weight Programmers" (NeurIPS 2021)☆47Updated last year
- ☆46Updated last month
- Fast training of unitary deep network layers from low-rank updates☆28Updated last year
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆57Updated last year
- Unofficial but Efficient Implementation of "Mamba: Linear-Time Sequence Modeling with Selective State Spaces" in JAX☆79Updated 9 months ago
- Recycling diverse models☆44Updated last year
- ☆16Updated last month
- Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount…☆49Updated last year
- PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)☆66Updated last year
- PyTorch implementation of Hinton's FF Algorithm with hard negatives sampling☆14Updated last year
- Stick-breaking attention☆34Updated last week