carloalbertobarbano / forward-forward-pytorchLinks
PyTorch implementation of Hinton's FF Algorithm with hard negatives sampling
☆15Updated 2 years ago
Alternatives and similar repositories for forward-forward-pytorch
Users that are interested in forward-forward-pytorch are comparing it to the libraries listed below
Sorting:
- ☆52Updated last year
- ☆75Updated 3 years ago
- Code for "Can We Scale Transformers to Predict Parameters of Diverse ImageNet Models?" [ICML 2023]☆37Updated last year
- ☆166Updated 2 years ago
- CIFAR10 ResNets implemented in JAX+Flax☆12Updated 3 years ago
- Reimplementation of Geoffrey Hinton's Forward-Forward Algorithm☆157Updated 2 years ago
- Code for papers Linear Algebra with Transformers (TMLR) and What is my Math Transformer Doing? (AI for Maths Workshop, Neurips 2022)☆76Updated last year
- ☆61Updated last year
- nanoGPT-like codebase for LLM training☆110Updated 2 weeks ago
- Replicating and dissecting the git-re-basin project in one-click-replication Colabs☆36Updated 3 years ago
- ☆56Updated last year
- Code release for REPAIR: REnormalizing Permuted Activations for Interpolation Repair☆51Updated last year
- ☆220Updated 2 years ago
- Implementation of Infini-Transformer in Pytorch☆113Updated 10 months ago
- Framework code with wandb, checkpointing, logging, configs, experimental protocols. Useful for fine-tuning models or training from scratc…☆152Updated 2 years ago
- Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts☆119Updated last year
- ☆51Updated last year
- CIFAR-10 speedruns: 94% in 2.6 seconds and 96% in 27 seconds☆326Updated last week
- ☆83Updated last year
- ☆82Updated last year
- A centralized place for deep thinking code and experiments☆87Updated 2 years ago
- FFCV-SSL Fast Forward Computer Vision for Self-Supervised Learning.☆211Updated 2 years ago
- Model Stock: All we need is just a few fine-tuned models☆127Updated 3 months ago
- Implementation of a memory efficient multi-head attention as proposed in the paper, "Self-attention Does Not Need O(n²) Memory"☆383Updated 2 years ago
- A repository for log-time feedforward networks☆223Updated last year
- Memory Efficient Attention (O(sqrt(n)) for Jax and PyTorch☆184Updated 2 years ago
- ☆209Updated 3 years ago
- This repository includes code to reproduce the tables in "Loss Landscapes are All You Need: Neural Network Generalization Can Be Explaine…☆40Updated 2 years ago
- Experiments around a simple idea for inducing multiple hierarchical predictive model within a GPT☆224Updated last year
- Implementation of the general framework for AMIE, from the paper "Towards Conversational Diagnostic AI", out of Google Deepmind☆70Updated last year