bytedance / ParaGen
ParaGen is a PyTorch deep learning framework for parallel sequence generation.
☆186Updated 2 years ago
Alternatives and similar repositories for ParaGen:
Users that are interested in ParaGen are comparing it to the libraries listed below
- ☆120Updated 3 years ago
- Implementation of "Glancing Transformer for Non-Autoregressive Neural Machine Translation"☆137Updated last year
- ☆166Updated 3 years ago
- A unified tokenization tool for Images, Chinese and English.☆151Updated last year
- 零样本学习测评基准,中文版☆54Updated 3 years ago
- Pretrain CPM-1☆51Updated 3 years ago
- Introduction to CPM☆163Updated 3 years ago
- Code for CPM-2 Pre-Train☆159Updated last year
- A paper list of pre-trained language models (PLMs).☆138Updated 3 years ago
- ☆53Updated 2 years ago
- NTK scaled version of ALiBi position encoding in Transformer.☆67Updated last year
- Ongoing research training transformer language models at scale, including: BERT & GPT-2☆69Updated last year
- Finetune CPM-1☆75Updated last year
- This is a code repository for the ACL 2022 paper "Learning to Generalize to More: Continuous Semantic Augmentation for Neural Machine Tra…☆52Updated 2 years ago
- 基于Gated Attention Unit的Transformer模型(尝鲜版)☆97Updated last year
- Finetune CPM-2☆83Updated last year
- NLU & NLG (zero-shot) depend on mengzi-t5-base-mt pretrained model☆75Updated 2 years ago
- A Multi-modal Model Chinese Spell Checker Released on ACL2021.☆155Updated last year
- RoFormer升级版☆150Updated 2 years ago
- ☆218Updated 2 years ago
- EVA: Large-scale Pre-trained Chit-Chat Models☆306Updated last year
- [ACL 2022] Structured Pruning Learns Compact and Accurate Models https://arxiv.org/abs/2204.00408☆192Updated last year
- reStructured Pre-training☆98Updated 2 years ago
- FLASHQuad_pytorch☆66Updated 2 years ago
- Tracking the progress in NLG for task-oriented dialogue system (resources, code, and new frontiers etc.)☆134Updated 2 years ago
- 中文图书语料MD5链接☆213Updated 11 months ago
- A more efficient GLM implementation!☆55Updated last year
- ⛵️The official PyTorch implementation for "BERT-of-Theseus: Compressing BERT by Progressive Module Replacing" (EMNLP 2020).☆310Updated last year
- A framework for cleaning Chinese dialog data☆264Updated 3 years ago
- 简洁易用版TinyBert:基于Bert进行知识蒸馏的预训练语言模型☆254Updated 4 years ago