godaai / scale-py-zhLinks
Python 数据科学加速:Dask、Ray、Xorbits、mpi4py
☆67Updated 4 months ago
Alternatives and similar repositories for scale-py-zh
Users that are interested in scale-py-zh are comparing it to the libraries listed below
Sorting:
- Triton Documentation in Chinese Simplified / Triton 中文文档☆71Updated 2 months ago
- LLM/MLOps/LLMOps☆94Updated last month
- ☆336Updated last week
- Course materials for MIT6.5940: TinyML and Efficient Deep Learning Computing☆47Updated 5 months ago
- A PyTorch-like deep learning framework. Just for fun.☆157Updated last year
- LLM101n: Let's build a Storyteller 中文版☆131Updated 10 months ago
- deep learning framework from scratch☆29Updated 3 years ago
- Code release for book "Efficient Training in PyTorch"☆69Updated 2 months ago
- Efficient, Flexible, and Highly Fault-Tolerant Model Service Management Based on SGLang☆53Updated 7 months ago
- 数据和计算科学知识库,分享机器学习系统、数据管理、数据科学原 创内容。☆66Updated 9 months ago
- UltraScale Playbook 中文版☆43Updated 3 months ago
- Easy, fast, and cheap pretrain,finetune, serving for everyone☆306Updated last week
- Materials for learning SGLang☆457Updated this week
- my cs notes☆51Updated 8 months ago
- ☆168Updated this week
- 看图学大模型☆311Updated 10 months ago
- 大模型/LLM推理和部署理论与实践☆278Updated 3 months ago
- ☆45Updated last year
- SJTU HPC 用户文档站点☆170Updated this week
- vLLM Documentation in Chinese Simplified / vLLM 中文文档☆80Updated last month
- Ray - A curated list of resources: https://github.com/ray-project/ray☆61Updated last week
- LLM notes, including model inference, transformer model structure, and llm framework code analysis notes.☆782Updated last week
- website☆429Updated 3 months ago
- [USENIX ATC '24] Accelerating the Training of Large Language Models using Efficient Activation Rematerialization and Optimal Hybrid Paral…☆57Updated 10 months ago
- ☆210Updated 2 years ago
- 通过带领大家解读Transformer模型来加深对模型的理解☆196Updated 3 weeks ago
- The repository has collected a batch of noteworthy MLSys bloggers (Algorithms/Systems)☆245Updated 5 months ago
- ☆87Updated 3 months ago
- A prefill & decode disaggregated LLM serving framework with shared GPU memory and fine-grained compute isolation.☆87Updated last month
- A tiny yet powerful LLM inference system tailored for researching purpose. vLLM-equivalent performance with only 2k lines of code (2% of …☆224Updated 2 weeks ago