SUSTech-IDEA / SUS-Chat
SUS-Chat: Instruction tuning done right
☆47Updated 9 months ago
Related projects ⓘ
Alternatives and complementary repositories for SUS-Chat
- XVERSE-65B: A multilingual large language model developed by XVERSE Technology Inc.☆132Updated 7 months ago
- Imitate OpenAI with Local Models☆85Updated 2 months ago
- 旨在对当前主流LLM进行一个直 观、具体、标准的评测☆92Updated last year
- deep learning☆148Updated 4 months ago
- 中文原生检索增强生成测评基准☆98Updated 6 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆121Updated 10 months ago
- the newest version of llama3,source code explained line by line using Chinese☆22Updated 6 months ago
- Mixture-of-Experts (MoE) Language Model☆180Updated 2 months ago
- The official codes for "Aurora: Activating chinese chat capability for Mixtral-8x7B sparse Mixture-of-Experts through Instruction-Tuning"☆257Updated 6 months ago
- 演示 vllm 对中文大语言模型的神奇效果☆31Updated last year
- 基于baichuan-7b的开源多模态大语言模型☆72Updated 11 months ago
- 文本去重☆67Updated 5 months ago
- qwen-7b and qwen-14b finetuning☆82Updated 6 months ago
- code for piccolo embedding model from SenseTime☆106Updated 5 months ago
- ☆90Updated 5 months ago
- 怎么训练一个LLM分词器☆129Updated last year
- ☆105Updated last year
- ☆129Updated 4 months ago
- Light local website for displaying performances from different chat models.☆85Updated 11 months ago
- ☆213Updated 5 months ago
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆52Updated 6 months ago
- Qwen-WisdomVast is a large model trained on 1 million high-quality Chinese multi-turn SFT data, 200,000 English multi-turn SFT data, and …☆18Updated 6 months ago
- zero零训练llm调参☆30Updated last year
- Another ChatGLM2 implementation for GPTQ quantization☆54Updated last year
- ☆72Updated 10 months ago
- code for Scaling Laws of RoPE-based Extrapolation☆70Updated last year
- 大语言模型指令调优工具(支持 FlashAttention)☆166Updated 10 months ago
- ☆125Updated last year
- Just for debug☆56Updated 8 months ago
- A more efficient GLM implementation!☆55Updated last year