gonglinyuan / StackingBERT
Source code for "Efficient Training of BERT by Progressively Stacking"
☆112Updated 5 years ago
Alternatives and similar repositories for StackingBERT
Users that are interested in StackingBERT are comparing it to the libraries listed below
Sorting:
- PyTorch implementation of Transformer-based Neural Machine Translation☆78Updated 2 years ago
- [EMNLP 2018] On Tree-Based Neural Sentence Modeling.☆64Updated 6 years ago
- A dual learning toolkit developed by Microsoft Research☆71Updated last year
- A Toolkit for Training, Tracking, Saving Models and Syncing Results☆61Updated 5 years ago
- souce code for "Accelerating Neural Transformer via an Average Attention Network"☆78Updated 5 years ago
- Reproduce the results of paper "Compressing Word Embeddings via Deep Compositional Code Learning" accepted ICLR 2018☆23Updated 6 years ago
- Non-autoregressive Neural Machine Translation (not a full version)☆71Updated 2 years ago
- PyTorch Language Model for 1-Billion Word (LM1B / GBW) Dataset☆123Updated 5 years ago
- Code for NIPS 2018 paper 'Frequency-Agnostic Word Representation'☆115Updated 6 years ago
- Phrase-Indexed Question Answering (PIQA)☆94Updated 6 years ago
- Efficient Contextualized Representation: Language Model Pruning for Sequence Labeling☆146Updated 5 years ago
- ☆74Updated 8 years ago
- Enhancing Sentence Embedding with Generalized Pooling☆11Updated 6 years ago
- A PyTorch implementation of Attention is all you need☆42Updated 6 years ago
- Code to reproduce results in our ACL 2018 paper "Did the Model Understand the Question?"☆33Updated 6 years ago
- Implementation of Densely Connected Attention Propagation for Reading Comprehension (NIPS 2018)☆69Updated 6 years ago
- Code from Jia and Liang, "Adversarial Examples for Evaluating Reading Comprehension Systems" (EMNLP 2017)☆118Updated 6 years ago
- This repo is not maintained. For latest version, please visit https://github.com/ictnlp. A collection of transformer's guides, implementa…☆44Updated 6 years ago
- LAMB Optimizer for Large Batch Training (TensorFlow version)☆120Updated 5 years ago
- Visualization for simple attention and Google's multi-head attention.☆67Updated 7 years ago
- Code for the paper "Are Sixteen Heads Really Better than One?"☆171Updated 5 years ago
- Code for the RecAdam paper: Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting.☆116Updated 4 years ago
- Method to improve inference time for BERT. This is an implementation of the paper titled "PoWER-BERT: Accelerating BERT Inference via Pro…☆61Updated last week
- ☆21Updated 5 years ago
- An Implementation of Bidirectional Attention Flow☆40Updated 7 years ago
- ☆47Updated 4 years ago
- Text Content Manipulation☆45Updated 4 years ago
- ☆119Updated 6 years ago
- [ACL'19] Code for "Semi-supervised Domain Adaptation for Dependency Parsing"☆15Updated 5 years ago
- The implementation of multi-branch attentive Transformer (MAT).☆33Updated 4 years ago