shiyemin / light-hf-proxyLinks
A light proxy solution for HuggingFace hub.
☆47Updated 2 years ago
Alternatives and similar repositories for light-hf-proxy
Users that are interested in light-hf-proxy are comparing it to the libraries listed below
Sorting:
- The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.☆68Updated 2 years ago
- ☆79Updated last year
- 基于baichuan-7b的开源多模态大语言模型☆72Updated last year
- Gaokao Benchmark for AI☆109Updated 3 years ago
- A Simple MLLM Surpassed QwenVL-Max with OpenSource Data Only in 14B LLM.☆38Updated last year
- ☆106Updated 2 years ago
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆41Updated last year
- XVERSE-65B: A multilingual large language model developed by XVERSE Technology Inc.☆141Updated last year
- The official codes for "Aurora: Activating chinese chat capability for Mixtral-8x7B sparse Mixture-of-Experts through Instruction-Tuning"☆266Updated last year
- GTS Engine: A powerful NLU Training System。GTS引擎(GTS-Engine)是一款开箱即用且性能强大的自然语言理解引擎,聚焦于小样本任务,能够仅用小样本就能自动化生产NLP模型。☆93Updated 2 years ago
- the newest version of llama3,source code explained line by line using Chinese☆22Updated last year
- SuperCLUE琅琊榜:中文通用大模型匿名对战评价基准☆145Updated last year
- Our 2nd-gen LMM☆34Updated last year
- Kanchil(鼷鹿)是世界上最小的偶蹄目动物,这个开源项目意在探索小模型(6B以下)是否也能具备和人类偏好对齐的能力。☆113Updated 2 years ago
- 大语言模型训练和服务调研☆36Updated 2 years ago
- Light local website for displaying performances from different chat models.☆87Updated 2 years ago
- Summarize all open source Large Languages Models and low-cost replication methods for Chatgpt.☆137Updated 2 years ago
- XVERSE-7B: A multilingual large language model developed by XVERSE Technology Inc.☆53Updated last year
- Multimodal chatbot with computer vision capabilities integrated, our 1st-gen LMM☆101Updated last year
- An opensource ChatBot built with ExpertPrompting which achieves 96% of ChatGPT's capability.☆299Updated 2 years ago
- XVERSE-MoE-A4.2B: A multilingual large language model developed by XVERSE Technology Inc.☆39Updated last year
- ChatGLM-6B-Slim:裁减掉20K图片Token的ChatGLM-6B,完全一样的性能,占用更小的显存。☆127Updated 2 years ago
- A large-scale language model for scientific domain, trained on redpajama arXiv split☆137Updated last year
- SUS-Chat: Instruction tuning done right☆49Updated last year
- 首个llama2 13b 中文版模型 (Base + 中文对话SFT,实现流畅多轮人机自然语言交互)☆91Updated 2 years ago
- The code and data for GrammarGPT.☆178Updated 2 years ago
- Imitate OpenAI with Local Models☆89Updated last year
- Another ChatGLM2 implementation for GPTQ quantization☆54Updated 2 years ago
- ☆235Updated last year
- code for Scaling Laws of RoPE-based Extrapolation☆73Updated 2 years ago