chunhuizhang / deeplearning-envsLinks
深度学习软硬件配置(小白向)
☆33Updated 3 weeks ago
Alternatives and similar repositories for deeplearning-envs
Users that are interested in deeplearning-envs are comparing it to the libraries listed below
Sorting:
- AFAC2024金融智能创新大赛☆52Updated 8 months ago
- the newest version of llama3,source code explained line by line using Chinese☆22Updated last year
- AGM阿格姆:AI基因图谱模型,从token-weight权重微粒角度,探索AI模型,GPT\LLM大模型的内在运作机制。☆29Updated 2 years ago
- accelerate generating vector by using onnx model☆18Updated last year
- Recursive Abstractive Processing for Tree-Organized Retrieval☆10Updated last year
- 模型压缩的小白入门教程☆22Updated last year
- 纯c++的全平台llm加速库,支持python调用,支持baichuan, glm, llama, moss基座,手机端流畅运行chatglm-6B级模型单卡可达10000+token / s,☆45Updated last year
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆41Updated last year
- ☆45Updated 7 months ago
- qwen models finetuning☆102Updated 5 months ago
- LLM RAG 应用,支持 API 调用,语音交互。☆11Updated last year
- 最少使用 3090 即可训练自己的比特大脑(miniLLM)🧠(进行中). Train your own BitBrain(A mini LLM) with just an RTX 3090 minimum.☆32Updated last month
- A repo for update and debug Mixtral-7x8B、MOE、ChatGLM3、LLaMa2、 BaChuan、Qwen an other LLM models include new models mixtral, mixtral 8x7b, …☆48Updated last week
- 大型语言模型实战指南:应用实践与场景落地☆76Updated 10 months ago
- Tutorial for Ray☆28Updated last year
- A more efficient GLM implementation!☆55Updated 2 years ago
- pytorch分布式训练☆67Updated 2 years ago
- 介绍docker、docker compose的使用。☆20Updated 11 months ago
- 千问14B和7B的逐行解释☆60Updated last year
- Qwen-WisdomVast is a large model trained on 1 million high-quality Chinese multi-turn SFT data, 200,000 English multi-turn SFT data, and …☆18Updated last year
- 中 文预训练ModernBert☆81Updated 4 months ago
- LoRA☆19Updated 2 years ago
- 大模型智能体Agent中文教程,博客代码仓库☆26Updated last month
- Efficient, Flexible, and Highly Fault-Tolerant Model Service Management Based on SGLang☆55Updated 9 months ago
- aigc evals☆11Updated last year
- (1)弹性区间标准化的旋转位置词嵌入编码器+peft LORA量化训练,提高万级tokens性能支持。(2) 证据理论解释学习,提升模型的复杂逻辑推理能力(3)兼容alpaca数据格式。☆45Updated 2 years ago
- The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.☆69Updated 2 years ago
- A MoE impl for PyTorch, [ATC'23] SmartMoE☆66Updated 2 years ago
- Imitate OpenAI with Local Models☆88Updated 11 months ago
- Delta-CoMe can achieve near loss-less 1-bit compressin which has been accepted by NeurIPS 2024☆56Updated 8 months ago