CyberZHG / keras-multi-head
A wrapper layer for stacking layers horizontally
☆228Updated 2 years ago
Alternatives and similar repositories for keras-multi-head:
Users that are interested in keras-multi-head are comparing it to the libraries listed below
- Attention mechanism for processing sequential data that considers the context for each timestamp.☆658Updated 2 years ago
- Transformer implemented in Keras☆372Updated 2 years ago
- Keras library for building (Universal) Transformers, facilitating BERT and GPT models☆535Updated 4 years ago
- A Keras+TensorFlow Implementation of the Transformer: Attention Is All You Need☆711Updated 3 years ago
- Keras Layer implementation of Attention for Sequential models☆444Updated last year
- Layer normalization implemented in Keras☆60Updated 2 years ago
- 自注意力与文本分类☆119Updated 6 years ago
- An example attention network with simple dataset.☆230Updated 5 years ago
- Implementations for a family of attention mechanisms, suitable for all kinds of natural language processing tasks and compatible with Ten…☆351Updated 11 months ago
- Re-implementation of ELMo on Keras☆135Updated last year
- attention-based LSTM/Dense implemented by Keras☆295Updated 6 years ago
- An Attention Layer in Keras☆43Updated 5 years ago
- Attention-based bidirectional LSTM for Classification Task (ICASSP)☆113Updated 2 years ago
- Implementation of the Transformer architecture described by Vaswani et al. in "Attention Is All You Need"☆28Updated 5 years ago
- TensorFlow implementation of focal loss☆186Updated 3 years ago
- document classification using LSTM + self attention☆113Updated 5 years ago
- Collection of custom layers and utility functions for Keras which are missing in the main framework.☆62Updated 4 years ago
- Binary and Categorical Focal loss implementation in Keras.☆278Updated 3 weeks ago
- Pytorch implementation of R-Transformer. Some parts of the code are adapted from the implementation of TCN and Transformer.☆226Updated 5 years ago
- Position embedding layers in Keras☆59Updated 2 years ago
- Keras implementation of BERT with pre-trained weights☆814Updated 5 years ago
- Transformer-based models implemented in tensorflow 2.x(using keras).☆75Updated 3 years ago
- An implementation of a sequence to sequence neural network using an encoder-decoder☆209Updated 5 years ago
- This repository contain various types of attention mechanism like Bahdanau , Soft attention , Additive Attention , Hierarchical Attention…☆125Updated 3 years ago
- 利用keras搭建的胶囊网络(capsule network文本分类模型,包含RNN、CNN、HAN等,其中keras_utils包含了capsule层和attention层的keras实现☆77Updated 6 years ago
- Multi-class metrics for Tensorflow☆225Updated 2 years ago
- Tensorflow Temporal Convolutional Network☆82Updated last year
- An LSTM in PyTorch with best practices (weight dropout, forget bias, etc.) built-in. Fully compatible with PyTorch LSTM.☆132Updated 5 years ago
- Visualizing RNNs using the attention mechanism☆750Updated 5 years ago
- How to use ELMo embeddings in Keras with Tensorflow Hub☆261Updated 6 years ago