google-research / growneuronLinks
☆55Updated 11 months ago
Alternatives and similar repositories for growneuron
Users that are interested in growneuron are comparing it to the libraries listed below
Sorting:
- ☆51Updated last year
- Code for "Can We Scale Transformers to Predict Parameters of Diverse ImageNet Models?" [ICML 2023]☆36Updated 10 months ago
- NEVIS'22: Benchmarking the next generation of never-ending learners☆102Updated 2 years ago
- Latest Weight Averaging (NeurIPS HITY 2022)☆30Updated 2 years ago
- [NeurIPS 2021] Code for Unsupervised Learning of Compositional Energy Concepts☆62Updated 2 years ago
- FID computation in Jax/Flax.☆28Updated last year
- ☆23Updated 2 years ago
- The official repository for our paper "The Dual Form of Neural Networks Revisited: Connecting Test Time Predictions to Training Patterns …☆16Updated last month
- Code repository for the paper "Meta-Learning via Classifier(-free) Diffusion Guidance"☆32Updated 2 years ago
- Implementation of Discrete Key / Value Bottleneck, in Pytorch☆88Updated 2 years ago
- [NeurIPS'20] Code for the Paper Compositional Visual Generation and Inference with Energy Based Models☆45Updated 2 years ago
- Implementation of a Transformer that Ponders, using the scheme from the PonderNet paper☆81Updated 3 years ago
- Repository for the PopulAtion Parameter Averaging (PAPA) paper☆26Updated last year
- Official code for the paper: "Metadata Archaeology"☆19Updated 2 years ago
- A GPT, made only of MLPs, in Jax☆58Updated 4 years ago
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixing☆50Updated 3 years ago
- ☆19Updated 3 years ago
- Official repository for the paper "Going Beyond Linear Transformers with Recurrent Fast Weight Programmers" (NeurIPS 2021)☆49Updated last month
- Recycling diverse models☆45Updated 2 years ago
- [NeurIPS 2022] DataMUX: Data Multiplexing for Neural Networks☆60Updated 2 years ago
- ☆41Updated 2 years ago
- ☆37Updated 11 months ago
- Implementation of "compositional attention" from MILA, a multi-head attention variant that is reframed as a two-step attention process wi…☆51Updated 3 years ago
- Open source code for paper "On the Learning and Learnability of Quasimetrics".☆32Updated 2 years ago
- ☆51Updated last year
- Implementation of Hierarchical Transformer Memory (HTM) for Pytorch☆75Updated 3 years ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆80Updated last year
- Replicating and dissecting the git-re-basin project in one-click-replication Colabs☆36Updated 2 years ago
- Official PyTorch Implementation of the Longhorn Deep State Space Model☆52Updated 7 months ago
- A centralized place for deep thinking code and experiments☆85Updated last year