guillaume-chevalier / Linear-Attention-Recurrent-Neural-Network
A recurrent attention module consisting of an LSTM cell which can query its own past cell states by the means of windowed multi-head attention. The formulas are derived from the BN-LSTM and the Transformer Network. The LARNN cell with attention can be easily used inside a loop on the cell state, just like any other RNN. (LARNN)
☆144Updated 6 years ago
Alternatives and similar repositories for Linear-Attention-Recurrent-Neural-Network:
Users that are interested in Linear-Attention-Recurrent-Neural-Network are comparing it to the libraries listed below
- Dilated RNNs in pytorch☆212Updated 5 years ago
- Implementation of IndRNN in Keras☆67Updated 4 years ago
- Collection of custom layers and utility functions for Keras which are missing in the main framework.☆62Updated 4 years ago
- Keras implementation of Nested LSTMs☆89Updated 6 years ago
- [ICLR'19] Trellis Networks for Sequence Modeling☆472Updated 5 years ago
- Repository for the ablation study of "Long Short-Term Memory Fully Convolutional Networks for Time Series Classification"☆54Updated 6 years ago
- Keras implementation of LSTM Variational Autoencoder☆227Updated 5 years ago
- Phased LSTM: Accelerating Recurrent Network Training for Long or Event-based Sequences (NIPS 2016) - Tensorflow 1.0☆128Updated 6 years ago
- PyTorch implementations of LSTM Variants (Dropout + Layer Norm)☆136Updated 3 years ago
- Tensorflow implementation for DilatedRNN☆347Updated 7 years ago
- Tensorflow implementation of a Hierarchical and Multiscale RNN, described in https://arxiv.org/abs/1609.01704☆135Updated 7 years ago
- Keras implementation of Attention Augmented Convolutional Neural Networks☆121Updated 5 years ago
- PTB Language Modelling task with LSTM + Attention layer☆31Updated 7 years ago
- A simple Tensorflow based library for deep and/or denoising AutoEncoder.☆149Updated 6 years ago
- Text autoencoder with LSTMs☆262Updated 5 years ago
- Miltiplicative LSTM for Keras 2.0+☆42Updated 7 years ago
- pytorch implementation of Independently Recurrent Neural Networks https://arxiv.org/abs/1803.04831☆121Updated 6 years ago
- Keras implementation of Phased LSTM [https://arxiv.org/abs/1610.09513]☆145Updated 4 years ago
- This repository contains the source for the paper "S-LSTM-GAN: Shared recurrent neural networks with adversarial training"☆88Updated 5 years ago
- Visualizing RNNs using the attention mechanism☆749Updated 5 years ago
- TensorFlow implementation of Independently Recurrent Neural Networks☆514Updated 3 years ago
- pytorch neural network attention mechanism☆147Updated 6 years ago
- STCN: Stochastic Temporal Convolutional Networks☆69Updated 4 years ago
- SRU implement in pytorch(Training RNNs as Fast as CNNs)☆46Updated 2 years ago
- Deep learning for time-series data☆49Updated 2 years ago
- Multi heads attention for image classification☆81Updated 6 years ago
- ☆11Updated 7 years ago
- Tools for loading standard data sets in machine learning☆203Updated 2 years ago
- An LSTM in PyTorch with best practices (weight dropout, forget bias, etc.) built-in. Fully compatible with PyTorch LSTM.☆132Updated 5 years ago
- Pytorch implementation of a basic language model using Attention in LSTM network☆26Updated 6 years ago