yitu-opensource / ConvBert
☆246Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for ConvBert
- ⛵️The official PyTorch implementation for "BERT-of-Theseus: Compressing BERT by Progressive Module Replacing" (EMNLP 2020).☆312Updated last year
- MPNet: Masked and Permuted Pre-training for Language Understanding https://arxiv.org/pdf/2004.09297.pdf☆286Updated 3 years ago
- ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators☆92Updated 3 years ago
- For the code release of our arXiv paper "Revisiting Few-sample BERT Fine-tuning" (https://arxiv.org/abs/2006.05987).☆184Updated last year
- The score code of FastBERT (ACL2020)☆604Updated 3 years ago
- pytorch implementation for Patient Knowledge Distillation for BERT Model Compression☆199Updated 5 years ago
- This is the official code repository for NumNet+(https://leaderboard.allenai.org/drop/submission/blu418v76glsbnh1qvd0)☆178Updated 3 months ago
- Transformer with Untied Positional Encoding (TUPE). Code of paper "Rethinking Positional Encoding in Language Pre-training". Improve exis…☆250Updated 3 years ago
- Pytorch Implementation of ALBERT(A Lite BERT for Self-supervised Learning of Language Representations)☆225Updated 3 years ago
- Codes for "TENER: Adapting Transformer Encoder for Named Entity Recognition"☆373Updated 4 years ago
- Code for the paper "Are Sixteen Heads Really Better than One?"☆168Updated 4 years ago
- Feel free to fine tune large BERT models with Multi-GPU and FP16 support.☆192Updated 4 years ago
- [ACL 2020] DeFormer: Decomposing Pre-trained Transformers for Faster Question Answering☆120Updated last year
- ☆201Updated last year
- ☆165Updated 2 years ago
- [ICLR 2020] Lite Transformer with Long-Short Range Attention☆599Updated 3 months ago
- Adversarial Training for Natural Language Understanding☆250Updated last year
- TensorFlow implementation of On the Sentence Embeddings from Pre-trained Language Models (EMNLP 2020)☆529Updated 3 years ago
- Code for the RecAdam paper: Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting.☆115Updated 3 years ago
- ☆212Updated 4 years ago
- ICML'2022: NLP From Scratch Without Large-Scale Pretraining: A Simple and Efficient Framework☆257Updated 10 months ago
- Code associated with the Don't Stop Pretraining ACL 2020 paper☆525Updated 2 years ago
- A masked language modeling objective to train a model to predict any subset of the target words, conditioned on both the input text and a…☆240Updated 3 years ago
- Repository for the paper "Optimal Subarchitecture Extraction for BERT"☆470Updated 2 years ago
- Few-shot Natural Language Generation for Task-Oriented Dialog☆190Updated last year
- An efficient implementation of the popular sequence models for text generation, summarization, and translation tasks. https://arxiv.org/p…☆433Updated 2 years ago
- The repo contains the code of the ACL2020 paper `Dice Loss for Data-imbalanced NLP Tasks`☆272Updated last year
- 对ACL2020 FastBERT论文的复现,论文地址//arxiv.org/pdf/2004.02178.pdf☆190Updated 2 years ago
- Multi-Task Deep Neural Networks for Natural Language Understanding☆157Updated last year
- multi-gpu pre-training in one machine for BERT from scratch without horovod (Data Parallelism)☆173Updated 3 weeks ago