xverse-ai / XVERSE-65B
XVERSE-65B: A multilingual large language model developed by XVERSE Technology Inc.
☆132Updated 7 months ago
Related projects ⓘ
Alternatives and complementary repositories for XVERSE-65B
- Imitate OpenAI with Local Models☆85Updated 2 months ago
- ☆125Updated last year
- ☆173Updated last year
- A high-throughput and memory-efficient inference and serving engine for LLMs☆124Updated 11 months ago
- The official codes for "Aurora: Activating chinese chat capability for Mixtral-8x7B sparse Mixture-of-Experts through Instruction-Tuning"☆257Updated 6 months ago
- ☆157Updated last year
- ☆213Updated 6 months ago
- Chinese large language model base generated through incremental pre-training on Chinese datasets☆234Updated last year
- SuperCLUE-Agent: 基于中文原生任务的Agent智能体核心能力测评基准☆78Updated last year
- ☆129Updated 4 months ago
- Light local website for displaying performances from different chat models.☆85Updated last year
- Mixture-of-Experts (MoE) Language Model☆180Updated 2 months ago
- SUS-Chat: Instruction tuning done right☆47Updated 10 months ago
- code for Scaling Laws of RoPE-based Extrapolation☆70Updated last year
- ☆158Updated last year
- 国内首个全参数训练的法律大模型 HanFei-1.0 (韩非)☆102Updated last year
- Just for debug☆56Updated 9 months ago
- 用于大模型 RLHF 进行人工数据标注排序的工具。A tool for manual response data annotation sorting in RLHF stage.☆243Updated last year
- Its an open source LLM based on MOE Structure.☆57Updated 4 months ago
- ☆181Updated this week
- deep learning☆149Updated 5 months ago