WangRongsheng / Aurora
The official codes for "Aurora: Activating chinese chat capability for Mixtral-8x7B sparse Mixture-of-Experts through Instruction-Tuning"
☆257Updated 6 months ago
Related projects ⓘ
Alternatives and complementary repositories for Aurora
- ☆213Updated 6 months ago
- BiLLa: A Bilingual LLaMA with Enhanced Reasoning Ability☆421Updated last year
- ☆173Updated last year
- 语言模型中文认知能力分析☆235Updated last year
- Light local website for displaying performances from different chat models.☆85Updated last year
- Mixture-of-Experts (MoE) Language Model☆180Updated 2 months ago
- deep learning☆149Updated 4 months ago
- XVERSE-65B: A multilingual large language model developed by XVERSE Technology Inc.☆132Updated 7 months ago
- ☆217Updated last year
- 旨在对当前主流LLM进行一个直观、具体、标准的评测☆92Updated last year
- 用于大模型 RLHF 进行人工数据标注排序的工具。A tool for manual response data annotation sorting in RLHF stage.☆242Updated last year
- Imitate OpenAI with Local Models☆85Updated 2 months ago
- Baichuan2代码的逐行解析版本,适合小白☆208Updated last year
- LongQLoRA: Extent Context Length of LLMs Efficiently☆159Updated last year
- ☆125Updated last year
- Firefly中文LLaMA-2大模型,支持增量预训练Baichuan2、Llama2、Llama、Falcon、Qwen、Baichuan、InternLM、Bloom等大模型☆397Updated last year
- ☆294Updated 4 months ago
- An opensource ChatBot built with ExpertPrompting which achieves 96% of ChatGPT's capability.☆296Updated last year
- ☆158Updated last year
- Chinese large language model base generated through incremental pre-training on Chinese datasets☆234Updated last year
- ☆187Updated this week
- ☆193Updated 6 months ago
- SUS-Chat: Instruction tuning done right☆47Updated 10 months ago
- 文本去重☆67Updated 5 months ago
- ☆157Updated last year
- 大语言模型指令调优工具(支持 FlashAttention)☆166Updated 10 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆123Updated 11 months ago
- Implement OpenAI APIs and plugin-enabled ChatGPT with open source LLM and other models.☆122Updated 5 months ago
- 更纯粹、更高压缩率的Tokenizer☆453Updated 7 months ago
- ☆297Updated last year