owenliang / qwen-dpo
通义千问的DPO训练
☆40Updated 6 months ago
Alternatives and similar repositories for qwen-dpo:
Users that are interested in qwen-dpo are comparing it to the libraries listed below
- 大语言模型应用:RAG、NL2SQL、聊天机器人、预训练、MOE混合专家模型、微调训练、强化学习、天池数据竞赛☆58Updated last month
- 本项目用于大模型数学解题能力方面的数据集合成,模型训练及评测,相关文章记录。☆80Updated 6 months ago
- ☆104Updated 8 months ago
- 使用单个24G显卡,从0开始训练LLM☆50Updated 4 months ago
- ☆36Updated 3 months ago
- pytorch分布式训练☆64Updated last year
- 训练一个对中文支持更好的LLaVA模型,并开源训练代码和数据。☆53Updated 6 months ago
- ☆73Updated 4 months ago
- ☆66Updated last year
- Qwen1.5-SFT(阿里, Ali), Qwen_Qwen1.5-2B-Chat/Qwen_Qwen1.5-7B-Chat微调(transformers)/LORA(peft)/推理☆54Updated 10 months ago
- ☆40Updated 7 months ago
- 阿里通义千问(Qwen-7B-Chat/Qwen-7B), 微调/LORA/推理☆85Updated 10 months ago
- ChatGLM-6B添加了RLHF的实现,以及 部分核心代码的逐行讲解 ,实例部分是做了个新闻短标题的生成,以及指定context推荐的RLHF的实现☆82Updated last year
- 阿里天池: 2023全球智能汽车AI挑战赛——赛道一:AI大模型检索问答 baseline 80+☆92Updated last year
- LLM Tokenizer with BPE algorithm☆30Updated 10 months ago
- qwen models finetuning☆92Updated last week
- 本项目用于Embedding模型的相关实验,包括Embedding模型评估、Embedding模型微调、Embedding模型量化等。☆44Updated 8 months ago
- simple decoder-only GTP model in pytorch☆36Updated 10 months ago
- 基于 LoRA 和 P-Tuning v2 的 ChatGLM-6B 高效参数微调☆54Updated last year
- 通用简单工具项目☆17Updated 5 months ago
- ☆44Updated 5 months ago
- baichuan LLM surpervised finetune by lora☆62Updated last year
- ☆67Updated 3 weeks ago
- 天池算法比赛《BetterMixture - 大模型数据混合挑战赛》的第一名top1解决方案☆27Updated 8 months ago
- 基于DPO算法微调语言大模型,简单好上手。☆31Updated 8 months ago
- Code for a New Loss for Mitigating the Bias of Learning Difficulties in Generative Language Models☆62Updated last month
- Qwen-WisdomVast is a large model trained on 1 million high-quality Chinese multi-turn SFT data, 200,000 English multi-turn SFT data, and …☆18Updated 11 months ago
- 大模型检索增强生成技术最佳实践。☆67Updated 6 months ago
- the newest version of llama3,source code explained line by line using Chinese☆22Updated 11 months ago