xiabingquan / distributed_pytorch_from_scratchLinks
PyTorch distributed training from scratch (for educational purposes only)
☆20Updated 8 months ago
Alternatives and similar repositories for distributed_pytorch_from_scratch
Users that are interested in distributed_pytorch_from_scratch are comparing it to the libraries listed below
Sorting:
- DeepSeek Native Sparse Attention pytorch implementation☆109Updated last month
- The repository has collected a batch of noteworthy MLSys bloggers (Algorithms/Systems)☆307Updated 11 months ago
- 注释的nano_vllm仓库,并且完成了MiniCPM4的适配以及注册新模型的功能☆118Updated 4 months ago
- ☆149Updated 5 months ago
- llm theoretical performance analysis tools and support params, flops, memory and latency analysis.☆113Updated 5 months ago
- ☆44Updated last year
- DLSlime: Flexible & Efficient Heterogeneous Transfer Toolkit☆84Updated this week
- Summary of some awesome work for optimizing LLM inference☆150Updated 3 weeks ago
- LLM training technologies developed by kwai☆66Updated 3 weeks ago
- Curated collection of papers in MoE model inference☆314Updated 2 months ago
- [DAC'25] Official implement of "HybriMoE: Hybrid CPU-GPU Scheduling and Cache Management for Efficient MoE Inference"☆94Updated 6 months ago
- [ICLR 2025] PEARL: Parallel Speculative Decoding with Adaptive Draft Length☆137Updated last month
- Since the emergence of chatGPT in 2022, the acceleration of Large Language Model has become increasingly important. Here is a list of pap…☆282Updated 9 months ago
- ☆107Updated 2 months ago
- Code release for book "Efficient Training in PyTorch"☆116Updated 8 months ago
- Build LLM from scratch☆70Updated last month
- Efficient Mixture of Experts for LLM Paper List☆149Updated 2 months ago
- High performance Transformer implementation in C++.☆146Updated 11 months ago
- ☆78Updated 3 weeks ago
- Omni_Infer is a suite of inference accelerators designed for the Ascend NPU platform, offering native support and an expanding feature se…☆92Updated last week
- [NeurIPS 2024] Efficient LLM Scheduling by Learning to Rank☆66Updated last year
- Implement some method of LLM KV Cache Sparsity☆41Updated last year
- UltraScale Playbook 中文版☆102Updated 9 months ago
- A tiny yet powerful LLM inference system tailored for researching purpose. vLLM-equivalent performance with only 2k lines of code (2% of …☆297Updated 6 months ago
- LLM Inference with Deep Learning Accelerator.☆56Updated 10 months ago
- A prefill & decode disaggregated LLM serving framework with shared GPU memory and fine-grained compute isolation.☆119Updated 7 months ago
- Puzzles for learning Triton, play it with minimal environment configuration!☆571Updated 3 weeks ago
- Bridge Megatron-Core to Hugging Face/Reinforcement Learning☆173Updated last week
- 🤖FFPA: Extend FlashAttention-2 with Split-D, ~O(1) SRAM complexity for large headdim, 1.8x~3x↑🎉 vs SDPA EA.☆239Updated last month
- A simple calculation for LLM MFU.☆50Updated 3 months ago