Hansen06 / GPT2-ChineseLinks
Chinese version of GPT2 training code, using BERT tokenizer.
☆37Updated 3 years ago
Alternatives and similar repositories for GPT2-Chinese
Users that are interested in GPT2-Chinese are comparing it to the libraries listed below
Sorting:
- TextGen: Implementation of Text Generation models, include LLaMA, BLOOM, GPT2, BART, T5, SongNet and so on. 文本生成模型,实现了包括LLaMA,ChatGLM,BLO…☆982Updated last year
- 人工精调的中文对话数据集和一段chatglm的微调代码☆1,195Updated 8 months ago
- PromptCLUE, 全中文任务支持零样本学习模型☆665Updated 2 years ago
- 骆驼:A Chinese finetuned instruction LLaMA. Developed by 陈启源 @ 华中师范大学 & 李鲁鲁 @ 商汤科技 & 冷子昂 @ 商汤科技☆718Updated 2 years ago
- pCLUE: 1000000+多任务提示学习数据集☆505Updated 3 years ago
- 使用peft库,对chatGLM-6B/chatGLM2-6B实现4bit的QLoRA高效微调,并做lora model和base model的merge及4bit的量化(quantize)。☆358Updated 2 years ago
- ChatGLM2-6B 全参数微调,支持多轮对话的高效微调。☆402Updated 2 years ago
- chatglm 6b finetuning and alpaca finetuning☆1,539Updated 10 months ago
- chatglm多gpu用deepspeed和☆409Updated last year
- 活字通用大模型☆391Updated last year
- 探索中文instruct数据在ChatGLM, LLaMA上的微调表现☆389Updated 2 years ago
- Tuning LLMs with no tears💦; Sample Design Engineering (SDE) for more efficient downstream-tuning.☆1,021Updated last year
- 开源SFT数据集整理,随时补充☆569Updated 2 years ago
- ChatGLM-6B 指令学习|指令数据|Instruct☆655Updated 2 years ago
- Easy and Efficient Finetuning LLMs. (Supported LLama, LLama2, LLama3, Qwen, Baichuan, GLM , Falcon) 大模型高效量化训练+部署.☆618Updated last year
- Code for fintune ChatGLM-6b using low-rank adaptation (LoRA)☆719Updated 2 years ago
- 基于ChatGLM-6B的中文问诊模型☆828Updated 2 years ago
- A Large-scale Chinese Short-Text Conversation Dataset and Chinese pre-training dialog models☆1,933Updated 2 years ago
- [COLING 2022] CSL: A Large-scale Chinese Scientific Literature Dataset 中文科学文献数据集☆661Updated 2 years ago
- Easy-to-use CPM for Chinese text generation(基于CPM的中文文本生成)☆531Updated 2 years ago
- alpaca中文指令微调数据集☆397Updated 2 years ago
- Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo☆1,093Updated last year
- ☆537Updated last year
- 夫子•明察司法大模型是由山东大学、浪潮云、中国政法大学联合研发,以 ChatGLM 为大模型底座,基于海量中文无监督司法语料与有监督司法微调数据训练的中文司法大模型。该模型支持法条检索、案例分析、三段论推理判决以及司法对话等功能,旨在为用户提供全方位、高精准的法律咨询与解答…☆366Updated 6 months ago
- 对ChatGLM直接使用RLHF提升或降低目标输出概率|Modify ChatGLM output with only RLHF☆197Updated 2 years ago
- unified embedding model☆878Updated 2 years ago
- "桃李“: 国际中文教育大模型☆189Updated 2 years ago
- PaddleNLP UIE模型的PyTorch版实现☆682Updated 2 years ago
- A Chinese medical ChatGPT based on LLaMa, training from large-scale pretrain corpus and multi-turn dialogue dataset.☆385Updated 2 years ago
- ⭐️ NLP Algorithms with transformers lib. Supporting Text-Classification, Text-Generation, Information-Extraction, Text-Matching, RLHF, SF…☆2,407Updated 2 years ago