carloalbertobarbano / forward-forward-pytorch
PyTorch implementation of Hinton's FF Algorithm with hard negatives sampling
☆14Updated last year
Related projects ⓘ
Alternatives and complementary repositories for forward-forward-pytorch
- ☆46Updated last month
- Code for "Can We Scale Transformers to Predict Parameters of Diverse ImageNet Models?" [ICML 2023]☆31Updated 2 months ago
- Replicating and dissecting the git-re-basin project in one-click-replication Colabs☆36Updated 2 years ago
- Code for papers Linear Algebra with Transformers (TMLR) and What is my Math Transformer Doing? (AI for Maths Workshop, Neurips 2022)☆64Updated 3 months ago
- ☆23Updated 5 months ago
- ☆58Updated 2 years ago
- ☆74Updated 11 months ago
- Reimplementation of Geoffrey Hinton's Forward-Forward Algorithm☆132Updated last year
- ☆51Updated 5 months ago
- A simple Python implementation of forward-forward NN training by G. Hinton from NeurIPS 2022☆21Updated last year
- Pytorch implementation of a simple way to enable (Stochastic) Frame Averaging for any network☆47Updated 3 months ago
- PyTorch implementation for "Long Horizon Temperature Scaling", ICML 2023☆19Updated last year
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆57Updated last year
- ☆76Updated 7 months ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆79Updated last year
- ☆73Updated 2 years ago
- ☆53Updated 3 months ago
- ☆161Updated last year
- MambaFormer in-context learning experiments and implementation for https://arxiv.org/abs/2402.04248☆34Updated 5 months ago
- A centralized place for deep thinking code and experiments☆77Updated last year
- Some personal experiments around routing tokens to different autoregressive attention, akin to mixture-of-experts☆108Updated last month
- Model Stock: All we need is just a few fine-tuned models☆92Updated last month
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆43Updated last year
- Exploring an idea where one forgets about efficiency and carries out attention across each edge of the nodes (tokens)☆43Updated last month
- A MAD laboratory to improve AI architecture designs 🧪☆95Updated 6 months ago
- Simple illustrative examples for energy-based models in PyTorch☆61Updated 4 years ago
- ☆20Updated 11 months ago
- ☆13Updated last year
- This repository includes code to reproduce the tables in "Loss Landscapes are All You Need: Neural Network Generalization Can Be Explaine…☆34Updated last year
- Implementation of Infini-Transformer in Pytorch☆104Updated last month