lonePatient / MobileBert_PyTorch
MobileBERT: a Compact Task-Agnostic BERT for Resource-Limited Devices
☆62Updated 4 years ago
Related projects: ⓘ
- pytorch implementation for Patient Knowledge Distillation for BERT Model Compression☆198Updated 5 years ago
- ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators☆91Updated 3 years ago
- R-Drop方法在中文任务上的简单实验☆90Updated 2 years ago
- 对ACL2020 FastBERT论文的复现,论文地址//arxiv.org/pdf/2004.02178.pdf☆191Updated 2 years ago
- ⛵️The official PyTorch implementation for "BERT-of-Theseus: Compressing BERT by Progressive Module Replacing" (EMNLP 2020).☆310Updated last year
- A general framework for knowledge distillation☆48Updated 3 years ago
- A PyTorch implementation of Google AI's BERT model provided with Google's pre-trained models, examples and utilities.☆68Updated 2 years ago
- 基于百度webqa与dureader数据集训练的Albert Large QA模型☆75Updated 4 years ago
- ☆50Updated last year
- ☆245Updated last year
- bert annotation, input and output for people from scratch, 代码注释, 有每一步的输入和输出, 适合初学者☆92Updated last year
- 离线端阅读理解应用 QA for mobile, Android & iPhone☆60Updated last year
- Finetune CPM-1☆75Updated last year
- A PyTorch implementation of Google AI's BERT model provided with Google's pre-trained models, examples and utilities.☆35Updated 5 years ago
- 论文模型复现☆42Updated 2 years ago
- bert-of-theseus via bert4keras☆31Updated 4 years ago
- ☆80Updated last year
- Knowledge Distillation from BERT☆51Updated 5 years ago
- 高性能小模型测评 Shared Tasks in NLPCC 2020. Task 1 - Light Pre-Training Chinese Language Model for NLP Task☆57Updated 4 years ago
- Code for the paper "Are Sixteen Heads Really Better than One?"☆165Updated 4 years ago
- tensorflow version of bert-of-theseus☆63Updated 3 years ago
- ☆44Updated 2 years ago
- [ACL 2022] Structured Pruning Learns Compact and Accurate Models https://arxiv.org/abs/2204.00408☆188Updated last year
- Unicoder model for understanding and generation.☆88Updated 9 months ago
- lightweighted deep learning inference service framework☆38Updated 3 years ago
- A PyTorch implementation of Transformer in "Attention is All You Need"☆103Updated 3 years ago
- reformer-pytorch中文版本,简单高效的生成模型。类似GPT2的效果☆16Updated last year
- Must-read papers on improving efficiency for pre-trained language models.☆100Updated last year
- Research code for ACL 2020 paper: "Distilling Knowledge Learned in BERT for Text Generation".☆129Updated 3 years ago
- ☆85Updated 4 years ago