shiyemin / light-hf-proxyLinks
A light proxy solution for HuggingFace hub.
☆46Updated last year
Alternatives and similar repositories for light-hf-proxy
Users that are interested in light-hf-proxy are comparing it to the libraries listed below
Sorting:
- The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.☆68Updated 2 years ago
- XVERSE-65B: A multilingual large language model developed by XVERSE Technology Inc.☆141Updated last year
- 基于baichuan-7b的开源多模态大语言模型☆72Updated last year
- the newest version of llama3,source code explained line by line using Chinese☆22Updated last year
- Gaokao Benchmark for AI☆108Updated 3 years ago
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆41Updated last year
- ☆106Updated 2 years ago
- XVERSE-7B: A multilingual large language model developed by XVERSE Technology Inc.☆53Updated last year
- Light local website for displaying performances from different chat models.☆87Updated last year
- ☆79Updated last year
- Another ChatGLM2 implementation for GPTQ quantization☆53Updated 2 years ago
- XVERSE-MoE-A4.2B: A multilingual large language model developed by XVERSE Technology Inc.☆39Updated last year
- A Simple MLLM Surpassed QwenVL-Max with OpenSource Data Only in 14B LLM.☆38Updated last year
- The official codes for "Aurora: Activating chinese chat capability for Mixtral-8x7B sparse Mixture-of-Experts through Instruction-Tuning"☆264Updated last year
- 百川Dynamic NTK-ALiBi的代码实现:无需微调即可推理更长文本☆49Updated 2 years ago
- Kanchil(鼷鹿)是世界上最小的偶蹄目动物,这个开源项目意在探索小模型(6B以下)是否也能具备和人类偏好对齐的能力。☆113Updated 2 years ago
- Summarize all open source Large Languages Models and low-cost replication methods for Chatgpt.☆137Updated 2 years ago
- GTS Engine: A powerful NLU Training System。GTS引擎(GTS-Engine)是一款开箱即用且性能强大的自然语言理解引擎,聚焦于小样本任务,能够仅用小样本就能自动化生产NLP模型。☆92Updated 2 years ago
- This is a personal reimplementation of Google's Infini-transformer, utilizing a small 2b model. The project includes both model and train…☆58Updated last year
- Our 2nd-gen LMM☆34Updated last year
- SUS-Chat: Instruction tuning done right☆49Updated last year
- This repository provides an implementation of "A Simple yet Effective Training-free Prompt-free Approach to Chinese Spelling Correction B…☆83Updated 3 months ago
- Mixture-of-Experts (MoE) Language Model☆189Updated last year
- Its an open source LLM based on MOE Structure.☆58Updated last year
- SkyScript-100M: 1,000,000,000 Pairs of Scripts and Shooting Scripts for Short Drama: https://arxiv.org/abs/2408.09333v2☆127Updated 11 months ago
- Delta-CoMe can achieve near loss-less 1-bit compressin which has been accepted by NeurIPS 2024☆57Updated 11 months ago
- 中文原生工业测评基准☆15Updated last year
- ChatGLM-6B-Slim:裁减掉20K图片Token的ChatGLM-6B,完全一样的性能,占用更小的显存。☆127Updated 2 years ago
- code for Scaling Laws of RoPE-based Extrapolation☆73Updated 2 years ago
- 百度QA100万数据集☆48Updated last year