taishan1994 / Chinese-LLaMA-Alpaca-LoRA-Tuning
使用LoRA对Chinese-LLaMA-Alpaca进行微调。
☆35Updated last year
Alternatives and similar repositories for Chinese-LLaMA-Alpaca-LoRA-Tuning:
Users that are interested in Chinese-LLaMA-Alpaca-LoRA-Tuning are comparing it to the libraries listed below
- deepspeed+trainer简单高效实现多卡微调大模型☆122Updated last year
- ChatGLM-6B添加了RLHF的实现,以及部分核心代码的逐行讲解 ,实例部分是做了个新闻短标题的生成,以及指定context推荐的RLHF的实现☆82Updated last year
- ☆102Updated 7 months ago
- baichuan LLM surpervised finetune by lora☆62Updated last year
- Large language Model fintuning bloom , opt , gpt, gpt2 ,llama,llama-2,cpmant and so on☆96Updated 9 months ago
- ☆64Updated last year
- llama,chatglm 等模型的微调☆85Updated 6 months ago
- BLOOM 模型的指令微调☆24Updated last year
- A full pipeline to finetune ChatGLM LLM with LoRA and RLHF on consumer hardware. Implementation of RLHF (Reinforcement Learning with Huma…☆134Updated last year
- 基于 LoRA 和 P-Tuning v2 的 ChatGLM-6B 高效参数微调☆54Updated last year
- 📔 对Chinese-LLaMA-Alpaca进行使用说明和核心代码注解☆50Updated last year
- 阿里天池: 2023全球智能汽车AI挑战赛——赛道一:AI大模型检索问答 baseline 80+☆86Updated last year
- 一套代码指令微调大模型☆38Updated last year
- basic framework for rag(retrieval augment generation)☆82Updated last year
- MEASURING MASSIVE MULTITASK CHINESE UNDERSTANDING☆88Updated 10 months ago
- 怎么训练一个LLM分词器☆138Updated last year
- A Massive Multi-Level Multi-Subject Knowledge Evaluation benchmark☆100Updated last year
- 使用 Qwen2ForSequenceClassification 简单实现文本分类任务。☆52Updated 8 months ago
- Baichuan-13B 指令微调☆89Updated last year
- Clustering and Ranking: Diversity-preserved Instruction Selection through Expert-aligned Quality Estimation☆74Updated 3 months ago
- ☆93Updated last year
- LLM+RAG for QA☆21Updated last year
- ChatGLM2-6B微调, SFT/LoRA, instruction finetune☆105Updated last year
- ☆159Updated last year
- LLM for NER☆61Updated 6 months ago
- Awesome Open-domain Dialogue Models,高质量开放域对话模型集合☆34Updated last year
- llama2 finetuning with deepspeed and lora☆172Updated last year
- 1st Solution For Conversational Multi-Doc QA Workshop & International Challenge @ WSDM'24 - Xiaohongshu.Inc☆163Updated 11 months ago
- 对ChatGLM直接使用RLHF提升或降低目标输出概率|Modify ChatGLM output with only RLHF☆192Updated last year
- 一个基于HuggingFace开发的大语言模型训练、测试工具。支持各模型的webui、终端预测,低参数量及全参数模型训练(预训练、SFT、RM、PPO、DPO)和融合、量化。☆208Updated last year