airaria / TextPruner
A PyTorch-based model pruning toolkit for pre-trained language models
☆385Updated last year
Alternatives and similar repositories for TextPruner:
Users that are interested in TextPruner are comparing it to the libraries listed below
- PromptBERT: Improving BERT Sentence Embeddings with Prompts☆334Updated last year
- text embedding☆144Updated last year
- A framework for cleaning Chinese dialog data☆269Updated 3 years ago
- A curated list of research papers in Sentence Reprsentation Learning and a sts leaderboard of sentence embeddings.☆315Updated last year
- ☆279Updated last year
- CPT: A Pre-Trained Unbalanced Transformer for Both Chinese Language Understanding and Generation☆488Updated 2 years ago
- Code used for sourcing and cleaning the BigScience ROOTS corpus☆309Updated 2 years ago
- [NeurIPS'22 Spotlight] A Contrastive Framework for Neural Text Generation☆471Updated last year
- Multi-language Enhanced LLaMA☆301Updated 2 years ago
- LERT: A Linguistically-motivated Pre-trained Language Model(语言学信息增强的预训练模型LERT)☆207Updated 2 years ago
- A novel method to tune language models. Codes and datasets for paper ``GPT understands, too''.☆930Updated 2 years ago
- The score code of FastBERT (ACL2020)☆604Updated 3 years ago
- ParaGen is a PyTorch deep learning framework for parallel sequence generation.☆186Updated 2 years ago
- EVA: Large-scale Pre-trained Chit-Chat Models☆307Updated 2 years ago
- 使用sentencepiece中BPE训练中文词表,并在transformers中进行使用。☆117Updated last year
- sentence-transformers to onnx 让sbert模型推理效率更快☆163Updated 3 years ago
- ☆308Updated 2 years ago
- SimBERT升级版(SimBERTv2)!☆441Updated 3 years ago
- 中文自然语言推理数据集(A large-scale Chinese Nature language inference and Semantic similarity calculation Dataset)☆429Updated 5 years ago
- Mengzi Pretrained Models☆536Updated 2 years ago
- The code for our paper "NSP-BERT: A Prompt-based Zero-Shot Learner Through an Original Pre-training Task —— Next Sentence Prediction"☆229Updated 2 years ago
- Implementation of Chinese ChatGPT☆287Updated last year
- ☆459Updated 10 months ago
- ICML'2022: NLP From Scratch Without Large-Scale Pretraining: A Simple and Efficient Framework☆258Updated last year
- A PyTorch-based knowledge distillation toolkit for natural language processing☆1,649Updated last year
- train llama on a single A100 80G node using 🤗 transformers and 🚀 Deepspeed Pipeline Parallelism☆218Updated last year
- 简洁易用版TinyBert:基于Bert进行知识蒸馏的预训练语言模型☆262Updated 4 years ago
- NLU & NLG (zero-shot) depend on mengzi-t5-base-mt pretrained model☆75Updated 2 years ago
- pCLUE: 1000000+多任务提示学习数据集☆491Updated 2 years ago
- Fine tune a T5 transformer model using PyTorch & Transformers🤗☆212Updated 4 years ago