intersun / PKD-for-BERT-Model-Compression
pytorch implementation for Patient Knowledge Distillation for BERT Model Compression
☆202Updated 5 years ago
Alternatives and similar repositories for PKD-for-BERT-Model-Compression
Users that are interested in PKD-for-BERT-Model-Compression are comparing it to the libraries listed below
Sorting:
- ⛵️The official PyTorch implementation for "BERT-of-Theseus: Compressing BERT by Progressive Module Replacing" (EMNLP 2020).☆312Updated last year
- Adversarial Training for Natural Language Understanding☆251Updated last year
- BERT distillation(基于BERT的蒸馏实验 )☆313Updated 4 years ago
- For the code release of our arXiv paper "Revisiting Few-sample BERT Fine-tuning" (https://arxiv.org/abs/2006.05987).☆184Updated last year
- Code for the paper "Are Sixteen Heads Really Better than One?"☆171Updated 5 years ago
- 对ACL2020 FastBERT论文的复现,论文地址//arxiv.org/pdf/2004.02178.pdf☆193Updated 3 years ago
- ☆78Updated 2 years ago
- ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators☆91Updated 3 years ago
- Code for the RecAdam paper: Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting.☆116Updated 4 years ago
- ☆50Updated last year
- Knowledge Distillation from BERT☆52Updated 6 years ago
- ☆121Updated 6 years ago
- UDA(Unsupervised Data Augmentation) implemented by pytorch☆276Updated 5 years ago
- A PyTorch implementation of "Improving BERT Fine-Tuning via Self-Ensemble and Self-Distillation"☆56Updated 5 years ago
- An attempt at replicating the Induction Network for FewRel data in Tensorflow☆177Updated 5 years ago
- Open source code for ACL 2020 Paper "Dynamic Fusion Network for Multi-Domain End-to-end Task-Oriented Dialog"☆104Updated 3 years ago
- Minghao Hu's thesis on Machine Reading Comprehension☆37Updated 5 years ago
- bert annotation, input and output for people from scratch, 代码注释, 有每一步的输入和输出, 适合初学者☆93Updated 2 years ago
- ☆251Updated 2 years ago
- ☆62Updated 5 years ago
- The score code of FastBERT (ACL2020)☆605Updated 3 years ago
- this is roberta wwm base distilled model which was distilled from roberta wwm by roberta wwm large☆65Updated 5 years ago
- Code for ACL2020 paper: Few-shot Slot Tagging with Collapsed Dependency Transfer and Label-enhanced Task-adaptive Projection Network☆153Updated 3 years ago
- ☆59Updated 5 years ago
- ☆93Updated 3 years ago
- Learning To Compare For Text , Few shot learning in text classification☆42Updated 4 years ago
- A list of recent papers about Meta / few-shot learning methods applied in NLP areas.☆231Updated 4 years ago
- The source code for the Cutoff data augmentation approach proposed in this paper: "A Simple but Tough-to-Beat Data Augmentation Approach …☆63Updated 4 years ago
- ☆167Updated 3 years ago
- Chinese GPT2: pre-training and fine-tuning framework for text generation☆188Updated 3 years ago