ml-jku / hopfield-boostingLinks
☆32Updated last year
Alternatives and similar repositories for hopfield-boosting
Users that are interested in hopfield-boosting are comparing it to the libraries listed below
Sorting:
- Code and pretrained models for the paper: "MatMamba: A Matryoshka State Space Model"☆61Updated 9 months ago
- ☆69Updated last year
- ☆82Updated last year
- ICLR 2025 - official implementation for "I-Con: A Unifying Framework for Representation Learning"☆111Updated 2 months ago
- Code for "Accelerating Training with Neuron Interaction and Nowcasting Networks" [to appear at ICLR 2025]☆20Updated 3 months ago
- Explorations into the proposal from the paper "Grokfast, Accelerated Grokking by Amplifying Slow Gradients"☆101Updated 8 months ago
- A Python Library for Learning Non-Euclidean Representations☆61Updated 3 weeks ago
- $100K or 100 Days: Trade-offs when Pre-Training with Academic Resources☆143Updated 3 months ago
- Source code for the paper "Positional Attention: Expressivity and Learnability of Algorithmic Computation"☆14Updated 3 months ago
- ☆56Updated 10 months ago
- The official repository for HyperZ⋅Z⋅W Operator Connects Slow-Fast Networks for Full Context Interaction.☆39Updated 4 months ago
- Sparse and discrete interpretability tool for neural networks☆63Updated last year
- Official implementation of MetaTree: Learning a Decision Tree Algorithm with Transformers☆113Updated 11 months ago
- Deep Networks Grok All the Time and Here is Why☆37Updated last year
- Implementation of 🌻 Mirasol, SOTA Multimodal Autoregressive model out of Google Deepmind, in Pytorch☆89Updated last year
- [ICML 2025] Roll the dice & look before you leap: Going beyond the creative limits of next-token prediction☆65Updated 3 months ago
- Implementation of Infini-Transformer in Pytorch☆111Updated 7 months ago
- ☆85Updated last year
- Repository for code used in the xVal paper☆142Updated last year
- Fork of Flame repo for training of some new stuff in development☆15Updated last month
- Official implementation of "BERTs are Generative In-Context Learners"☆32Updated 5 months ago
- Official Repository of Pretraining Without Attention (BiGS), BiGS is the first model to achieve BERT-level transfer learning on the GLUE …☆114Updated last year
- Understanding how features learned by neural networks evolve throughout training☆36Updated 10 months ago
- Automatic Integration for Neural Spatio-Temporal Point Process models (AI-STPP) is a new paradigm for exact, efficient, non-parametric inf…☆25Updated 10 months ago
- LoRA-Ensemble: Efficient Uncertainty Modelling for Self-attention Networks☆48Updated 10 months ago
- Implementation of Spectral State Space Models☆16Updated last year
- Quantification of Uncertainty with Adversarial Models☆30Updated 2 years ago
- ☆41Updated last year
- Explorations into the recently proposed Taylor Series Linear Attention☆100Updated last year
- Implementation of the general framework for AMIE, from the paper "Towards Conversational Diagnostic AI", out of Google Deepmind☆67Updated 11 months ago