mcbal / deep-implicit-attentionLinks
Implementation of deep implicit attention in PyTorch
☆65Updated 4 years ago
Alternatives and similar repositories for deep-implicit-attention
Users that are interested in deep-implicit-attention are comparing it to the libraries listed below
Sorting:
- Meta-learning inductive biases in the form of useful conserved quantities.☆39Updated 3 years ago
- Tensorflow implementation and notebooks for Implicit Maximum Likelihood Estimation☆67Updated 3 years ago
- Usable implementation of Emerging Symbol Binding Network (ESBN), in Pytorch☆25Updated 5 years ago
- ☆50Updated 5 years ago
- Neural Turing Machines in pytorch☆49Updated 4 years ago
- Official code repository of the paper Linear Transformers Are Secretly Fast Weight Programmers.☆111Updated 4 years ago
- Very deep VAEs in JAX/Flax☆46Updated 4 years ago
- ICML 2020 Paper: Latent Variable Modelling with Hyperbolic Normalizing Flows☆54Updated 3 years ago
- Pytorch implementation of the Power Spherical distribution☆75Updated last year
- Code for the article "What if Neural Networks had SVDs?", to be presented as a spotlight paper at NeurIPS 2020.☆77Updated last year
- Official code for UnICORNN (ICML 2021)☆28Updated 4 years ago
- code for "Semi-Discrete Normalizing Flows through Differentiable Tessellation"☆26Updated 3 years ago
- Experiments for Meta-Learning Symmetries by Reparameterization☆58Updated 4 years ago
- Efficient Householder Transformation in PyTorch☆69Updated 4 years ago
- [NeurIPS 2020] Neural Manifold Ordinary Differential Equations (https://arxiv.org/abs/2006.10254)☆125Updated 2 years ago
- JAX exponential map normalising flows on sphere☆17Updated 5 years ago
- General Invertible Transformations for Flow-based Generative Models☆18Updated 5 years ago
- Riemannian Convex Potential Maps☆67Updated 2 years ago
- ☆64Updated 2 years ago
- Stochastic Automatic Differentiation library for PyTorch.☆208Updated last year
- Official repository for the paper "Going Beyond Linear Transformers with Recurrent Fast Weight Programmers" (NeurIPS 2021)☆51Updated 7 months ago
- A public repository for our paper, Rao-Blackwellized Stochastic Gradients for Discrete Distributions☆22Updated 6 years ago
- An implementation of (Induced) Set Attention Block, from the Set Transformers paper☆67Updated 3 years ago
- [ICML'21 Oral] Improving Lossless Compression Rates via Monte Carlo Bits-Back Coding☆14Updated 4 years ago
- ☆100Updated 4 years ago
- Structured matrices for compressing neural networks☆67Updated 2 years ago
- The official repository for our paper "The Devil is in the Detail: Simple Tricks Improve Systematic Generalization of Transformers". We s…☆67Updated 3 years ago
- Transformers with doubly stochastic attention☆53Updated 3 years ago
- Codebase for Learning Invariances in Neural Networks☆96Updated 3 years ago
- Humans understand novel sentences by composing meanings and roles of core language components. In contrast, neural network models for nat…☆27Updated 5 years ago