Tencent-Hunyuan / Hunyuan-7BLinks
Tencent Hunyuan 7B (short as Hunyuan-7B) is one of the large language dense models of Tencent Hunyuan
☆66Updated 2 months ago
Alternatives and similar repositories for Hunyuan-7B
Users that are interested in Hunyuan-7B are comparing it to the libraries listed below
Sorting:
- XVERSE-MoE-A36B: A multilingual large language model developed by XVERSE Technology Inc.☆38Updated last year
- ☆180Updated last month
- Fused Qwen3 MoE layer for faster training, compatible with HF Transformers, LoRA, 4-bit quant, Unsloth☆197Updated this week
- xllamacpp - a Python wrapper of llama.cpp☆60Updated last week
- Model compression toolkit engineered for enhanced usability, comprehensiveness, and efficiency.☆185Updated this week
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆41Updated last year
- GLM Series Edge Models☆149Updated 4 months ago
- ☆49Updated 8 months ago
- XVERSE-MoE-A4.2B: A multilingual large language model developed by XVERSE Technology Inc.☆39Updated last year
- ☆102Updated last year
- ☆89Updated 5 months ago
- Its an open source LLM based on MOE Structure.☆58Updated last year
- A repository aimed at pruning DeepSeek V3, R1 and R1-zero to a usable size☆74Updated last month
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆137Updated last year
- Lightweight toolkit package to train and fine-tune 1.58bit Language models☆92Updated 5 months ago
- ☆51Updated last year
- Family of instruction-following LLMs powered by Evol-Instruct: WizardLM, WizardCoder☆44Updated last year
- A fast RWKV Tokenizer written in Rust☆53Updated 2 months ago
- ☆57Updated 8 months ago
- ☆97Updated 2 months ago
- Code for KaLM-Embedding models☆93Updated 3 months ago
- ☆21Updated last year
- A light proxy solution for HuggingFace hub.☆46Updated last year
- Deep Reasoning Translation (DRT) Project☆236Updated last month
- Auto Thinking Mode switch for Qwen3 in Open webui☆68Updated 5 months ago
- Delta-CoMe can achieve near loss-less 1-bit compressin which has been accepted by NeurIPS 2024☆57Updated 11 months ago
- ☆298Updated 4 months ago
- ☆55Updated 11 months ago
- The official repo for “Unleashing the Reasoning Potential of Pre-trained LLMs by Critique Fine-Tuning on One Problem” [EMNLP25]☆32Updated last month
- Data preparation code for CrystalCoder 7B LLM☆45Updated last year