JacksonWuxs / Forward-Forward-Network
Implementation of Forward Forward Network proposed by Hinton in NIPS 2022.
☆169Updated 2 years ago
Alternatives and similar repositories for Forward-Forward-Network
Users that are interested in Forward-Forward-Network are comparing it to the libraries listed below
Sorting:
- Reimplementation of Geoffrey Hinton's Forward-Forward Algorithm☆148Updated last year
- ☆205Updated 2 years ago
- ☆62Updated 3 years ago
- ☆190Updated last year
- Implementation of Block Recurrent Transformer - Pytorch☆217Updated 8 months ago
- A curated list of Model Merging methods.☆92Updated 7 months ago
- PyTorch implementation of Mixer-nano (#parameters is 0.67M, originally Mixer-S/16 has 18M) with 90.83 % acc. on CIFAR-10. Training from s…☆32Updated 3 years ago
- Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorch☆287Updated last month
- PyTorch implementation of "From Sparse to Soft Mixtures of Experts"☆56Updated last year
- A simple Python implementation of forward-forward NN training by G. Hinton from NeurIPS 2022☆21Updated 2 years ago
- Demonstrations of Loss of Plasticity and Implementation of Continual Backpropagation☆301Updated 6 months ago
- ☆127Updated last year
- ☆229Updated last year
- Implementation of "Gradients without backpropagation" paper (https://arxiv.org/abs/2202.08587) using functorch☆109Updated last year
- A repository for DenseSSMs☆87Updated last year
- PyTorch implementation for Vision Transformer[Dosovitskiy, A.(ICLR'21)] modified to obtain over 90% accuracy FROM SCRATCH on CIFAR-10 wit…☆191Updated last year
- ☆64Updated 8 months ago
- The official PyTorch implementation of the paper: Xili Dai, Shengbang Tong, et al. "Closed-Loop Data Transcription to an LDR via Minimaxi…☆62Updated 2 years ago
- [EMNLP 2023, Main Conference] Sparse Low-rank Adaptation of Pre-trained Language Models☆76Updated last year
- A pytorch implementation of Cascaded Forward (CaFo) Algorithm☆22Updated 2 years ago
- Recurrent Memory Transformer☆149Updated last year
- Implementation of ASAM: Adaptive Sharpness-Aware Minimization for Scale-Invariant Learning of Deep Neural Networks, ICML 2021.☆143Updated 3 years ago
- [ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Di…☆57Updated 7 months ago
- ☆83Updated 9 months ago
- ☆146Updated last year
- PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)☆71Updated last year
- Code for paper "Diffusion Language Models Can Perform Many Tasks with Scaling and Instruction-Finetuning"☆77Updated last year
- Code used in "Understanding Dimensional Collapse in Contrastive Self-supervised Learning" paper.☆77Updated 2 years ago
- Inference Speed Benchmark for Learning to (Learn at Test Time): RNNs with Expressive Hidden States☆66Updated 9 months ago
- Sequence modeling with Mega.☆295Updated 2 years ago