使用qlora对中文大语言模型进行微调,包含ChatGLM、Chinese-LLaMA-Alpaca、BELLE
☆89Jun 27, 2023Updated 2 years ago
Alternatives and similar repositories for qlora-chinese-LLM
Users that are interested in qlora-chinese-LLM are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- 使用peft库,对chatGLM-6B/chatGLM2-6B实现4bit的QLoRA高效微调,并做lora model和base model的merge及4bit的量化(quantize)。☆360Aug 22, 2023Updated 2 years ago
- 面向金融领域的小样本跨类迁移事件抽取 第三名 方案及代码☆17Dec 23, 2020Updated 5 years ago
- moss chat finetuning☆51Apr 23, 2024Updated last year
- Easy and Efficient Finetuning LLMs. (Supported LLama, LLama2, LLama3, Qwen, Baichuan, GLM , Falcon) 大模型高效量化训练+部署.☆621Jan 24, 2025Updated last year
- Source code of ACL 2023 accepted paper "AD-KD: Attribution-Driven Knowledge Distillation for Language Model Compression"☆13Jun 14, 2023Updated 2 years ago
- Managed Database hosting by DigitalOcean • AdPostgreSQL, MySQL, MongoDB, Kafka, Valkey, and OpenSearch available. Automatically scale up storage and focus on building your apps.
- 中文文本的向量表示方法(Sentence-BERT, CoSENT)的PyTorch简单实现,可以用于文本相似度计算。☆10Mar 27, 2022Updated 4 years ago
- ☆11May 2, 2023Updated 2 years ago
- 2021搜狐校园文本匹配算法大赛☆16Jun 4, 2021Updated 4 years ago
- 基于规则匹配的问答系统中的解析器,the parser of based rule QA system☆12Mar 13, 2020Updated 6 years ago
- aigc evals☆10Dec 2, 2023Updated 2 years ago
- Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调☆3,723Oct 12, 2023Updated 2 years ago
- 基于 LoRA 和 P-Tuning v2 的 ChatGLM-6B 高效参数微调☆55May 17, 2023Updated 2 years ago
- 大语言模型微调的项目,包含了使用QLora微调ChatGLM和LLama☆29Jun 26, 2023Updated 2 years ago
- 大语言模型指令调优工具(支持 FlashAttention)☆176Jan 4, 2024Updated 2 years ago
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting with the flexibility to host WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Cloudways by DigitalOcean.
- A large-scale 7B pretraining language model developed by BaiChuan-Inc.☆5,676Jul 18, 2024Updated last year
- 本项目采用Keras和ALBERT实现文本多标签分类任务,其中对ALBERT进行微调。☆13Jan 5, 2021Updated 5 years ago
- BiLLa: A Bilingual LLaMA with Enhanced Reasoning Ability☆416Jun 1, 2023Updated 2 years ago
- DataFountain第五届达观杯第4名方案☆11Dec 3, 2021Updated 4 years ago
- 🌈 NERpy: Implementation of Named Entity Recognition using Python. 命名实体识别工具 ,支持BertSoftmax、BertSpan等模型,开箱即用。☆118Feb 19, 2024Updated 2 years ago
- chatglm 6b finetuning and alpaca finetuning☆1,536Mar 9, 2025Updated last year
- ☆13May 25, 2023Updated 2 years ago
- 文本去重☆78May 23, 2024Updated last year
- 百亿参数的中英文双语基座大模型☆2,412Jul 28, 2023Updated 2 years ago
- Simple, predictable pricing with DigitalOcean hosting • AdAlways know what you'll pay with monthly caps and flat pricing. Enterprise-grade infrastructure trusted by 600k+ customers.
- using lear to do ner extraction☆29Mar 13, 2022Updated 4 years ago
- 国内首个全参数训练的法律大模型 HanFei-1.0 (韩非)☆130Oct 28, 2023Updated 2 years ago
- Source code for ACL 2023 paper Decoder Tuning: Efficient Language Understanding as Decoding☆52Jun 25, 2023Updated 2 years ago
- Deepspeed、LLM、Medical_Dialogue、医疗大模型、预训练、微调☆297Jun 7, 2024Updated last year
- GTS Engine: A powerful NLU Training System。GTS引擎(GTS-Engine)是一款开箱即用且性能强大的自然语言理解引擎,聚焦于小样本任务,能够仅用小样本就能自动化生产NLP模型。☆93Feb 28, 2023Updated 3 years ago
- 抽取式NLP模型(阅读理解模型,MRC)实现词义消歧(WSD)☆14May 10, 2022Updated 3 years ago
- huggingface ChineseBert Tokenizer☆17Apr 16, 2022Updated 3 years ago
- 中文原生工业测评基准☆15Mar 21, 2024Updated 2 years ago
- 采用bert进行事件抽取,[cls]进行事件分类,最后一层向量进行序列标注,两个任务同时训练。☆13Jun 7, 2021Updated 4 years ago
- End-to-end encrypted email - Proton Mail • AdSpecial offer: 40% Off Yearly / 80% Off First Month. All Proton services are open source and independently audited for security.
- QLoRA: Efficient Finetuning of Quantized LLMs☆10,865Jun 10, 2024Updated last year
- DeepEE: Deep Event Extraction Algorithm Gallery (基于深度学习的开源中文事件抽取算法汇总)☆43Dec 11, 2022Updated 3 years ago
- 基于ChatGLM-6B、ChatGLM2-6B、ChatGLM3-6B模型,进行下游具体任务微调,涉及Freeze、Lora、P-tuning、全参微调等☆2,781Dec 12, 2023Updated 2 years ago
- 探索中文instruct数据在ChatGLM, LLaMA上的微调表现☆389Apr 4, 2023Updated 3 years ago
- Train large COMET (T5-3B/GPT2-XL) with small memory (on 11GB memory GPUs like 1080/2080) using DeepSpeed.☆14Jan 23, 2022Updated 4 years ago
- ChatGLM-6B 指令学习|指令数据|Instruct☆653Apr 10, 2023Updated 3 years ago
- The official codes for "Aurora: Activating chinese chat capability for Mixtral-8x7B sparse Mixture-of-Experts through Instruction-Tuning"☆263May 9, 2024Updated last year