YunwenTechnology / Chinese-Data-Distill-From-R1Links
中文基于满血DeepSeek-R1蒸馏数据集
☆56Updated 3 months ago
Alternatives and similar repositories for Chinese-Data-Distill-From-R1
Users that are interested in Chinese-Data-Distill-From-R1 are comparing it to the libraries listed below
Sorting:
- 怎么训练一个LLM分词器☆149Updated last year
- Imitate OpenAI with Local Models☆87Updated 9 months ago
- ☆228Updated last year
- 大语言模型指令调优工具(支持 FlashAttention)☆173Updated last year
- 专注于中文领域大语言模型,落地到某个行业某个领域,成为一个行业大模型、公司级别或行业级别领域大模型。☆118Updated 3 months ago
- ☆169Updated last year
- ☆97Updated last year
- ☆63Updated 2 years ago
- 中文原生检索增强生成测评基准☆118Updated last year
- Official Repository for SIGIR2024 Demo Paper "An Integrated Data Processing Framework for Pretraining Foundation Models"☆80Updated 9 months ago
- Alpaca Chinese Dataset -- 中文指令微调数据集☆205Updated 8 months ago
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆56Updated last year
- This repository provides an implementation of the paper "A Simple yet Effective Training-free Prompt-free Approach to Chinese Spelling Co…☆69Updated 2 months ago
- SuperCLUE-Agent: 基于中文原生任务的Agent智能体核心能力测评基准☆88Updated last year
- ☆142Updated 11 months ago
- baichuan LLM surpervised finetune by lora☆63Updated last year
- Dataset and evaluation script for "Evaluating Hallucinations in Chinese Large Language Models"☆128Updated last year
- ☆162Updated 2 years ago
- A demo built on Megrez-3B-Instruct, integrating a web search tool to enhance the model's question-and-answer capabilities.☆38Updated 5 months ago
- 中文 Instruction tuning datasets☆131Updated last year
- 使用单个24G显卡,从0开始训练LLM☆54Updated last week
- 阿里通义千问(Qwen-7B-Chat/Qwen-7B), 微调/LORA/推理☆104Updated last year
- ☆69Updated last year
- ☆144Updated last year
- CLongEval: A Chinese Benchmark for Evaluating Long-Context Large Language Models☆40Updated last year
- ChatGLM2-6B微调, SFT/LoRA, instruction finetune☆108Updated last year
- 基于DPO算法微调语言大模型,简单好上手。☆39Updated 11 months ago
- 基于ChatGLM2-6B进行微调,包括全参数、参数有效性、量化感知训练等,可实现指令微调、多轮对话微调等。☆25Updated last year
- 语言模型中文认知能力分析☆236Updated last year
- ☆141Updated last year