Tencent-Hunyuan / Hunyuan-7BLinks
Tencent Hunyuan 7B (short as Hunyuan-7B) is one of the large language dense models of Tencent Hunyuan
☆41Updated this week
Alternatives and similar repositories for Hunyuan-7B
Users that are interested in Hunyuan-7B are comparing it to the libraries listed below
Sorting:
- XVERSE-MoE-A36B: A multilingual large language model developed by XVERSE Technology Inc.☆39Updated 10 months ago
- ☆33Updated this week
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆40Updated last year
- This is an NVIDIA AI Workbench example project that demonstrates an end-to-end model development workflow using Llamafactory.☆63Updated 9 months ago
- Code for KaLM-Embedding models☆86Updated last month
- ☆288Updated 2 months ago
- Mixture-of-Experts (MoE) Language Model☆189Updated 11 months ago
- Fused Qwen3 MoE layer for faster training, compatible with HF Transformers, LoRA, 4-bit quant, Unsloth☆142Updated last week
- GLM Series Edge Models☆146Updated last month
- Implemented a script that automatically adjusts Qwen3's inference and non-inference capabilities, based on an OpenAI-like API. The infere…☆22Updated 3 months ago
- ☆51Updated last year
- Submodular optimization for context engineering: query fan-out, text selection, passage reranking☆67Updated 3 weeks ago
- A light proxy solution for HuggingFace hub.☆46Updated last year
- GPT-4 Level Conversational QA Trained In a Few Hours☆63Updated 11 months ago
- Official repository for RAGViz: Diagnose and Visualize Retrieval-Augmented Generation [EMNLP 2024]☆86Updated 6 months ago
- vLLM adapter for a TGIS-compatible gRPC server.☆34Updated this week
- ☆90Updated 2 months ago
- ☆102Updated 11 months ago
- XVERSE-MoE-A4.2B: A multilingual large language model developed by XVERSE Technology Inc.☆40Updated last year
- the newest version of llama3,source code explained line by line using Chinese☆22Updated last year
- Data preparation code for CrystalCoder 7B LLM☆45Updated last year
- CPM.cu is a lightweight, high-performance CUDA implementation for LLMs, optimized for end-device inference and featuring cutting-edge tec…☆169Updated last week
- ☆111Updated last year
- Its an open source LLM based on MOE Structure.☆58Updated last year
- Deep Reasoning Translation (DRT) Project☆227Updated 2 months ago
- GGML implementation of BERT model with Python bindings and quantization.☆56Updated last year
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆136Updated last year
- Inference Llama 2 in C++☆43Updated last year
- Evaluation of bm42 sparse indexing algorithm☆68Updated last year
- ☆48Updated 6 months ago