stepfun-ai / StepMeshLinks
☆342Updated this week
Alternatives and similar repositories for StepMesh
Users that are interested in StepMesh are comparing it to the libraries listed below
Sorting:
- A lightweight design for computation-communication overlap.☆213Updated last week
- High performance Transformer implementation in C++.☆148Updated last year
- nnScaler: Compiling DNN models for Parallel Training☆124Updated 4 months ago
- Dynamic Memory Management for Serving LLMs without PagedAttention☆457Updated 8 months ago
- Allow torch tensor memory to be released and resumed later☆213Updated 2 weeks ago
- DeepSeek-V3/R1 inference performance simulator☆176Updated 10 months ago
- Pipeline Parallelism Emulation and Visualization☆77Updated 3 weeks ago
- NVSHMEM‑Tutorial: Build a DeepEP‑like GPU Buffer☆158Updated 4 months ago
- A low-latency & high-throughput serving engine for LLMs☆471Updated 3 weeks ago
- Zero Bubble Pipeline Parallelism☆449Updated 8 months ago
- ☆130Updated last year
- MSCCL++: A GPU-driven communication stack for scalable AI applications☆455Updated last week
- LLM training technologies developed by kwai☆70Updated last week
- NVIDIA NVSHMEM is a parallel programming interface for NVIDIA GPUs based on OpenSHMEM. NVSHMEM can significantly reduce multi-process com…☆459Updated last month
- Perplexity GPU Kernels☆554Updated 2 months ago
- Distributed MoE in a Single Kernel [NeurIPS '25]☆188Updated last week
- A NCCL extension library, designed to efficiently offload GPU memory allocated by the NCCL communication library.☆87Updated last month
- Accelerating MoE with IO and Tile-aware Optimizations☆563Updated last week
- Utility scripts for PyTorch (e.g. Make Perfetto show some disappearing kernels, Memory profiler that understands more low-level allocatio…☆82Updated 4 months ago
- A tiny yet powerful LLM inference system tailored for researching purpose. vLLM-equivalent performance with only 2k lines of code (2% of …☆312Updated 7 months ago
- A prefill & decode disaggregated LLM serving framework with shared GPU memory and fine-grained compute isolation.☆123Updated last month
- DeepXTrace is a lightweight tool for precisely diagnosing slow ranks in DeepEP-based environments.☆91Updated 2 weeks ago
- ☆158Updated last year
- High-performance distributed data shuffling (all-to-all) library for MoE training and inference☆109Updated last month
- ☆83Updated 3 months ago
- PyTorch distributed training acceleration framework☆55Updated 5 months ago
- Since the emergence of chatGPT in 2022, the acceleration of Large Language Model has become increasingly important. Here is a list of pap…☆283Updated 10 months ago
- ☆105Updated last year
- ByteCheckpoint: An Unified Checkpointing Library for LFMs☆268Updated last month
- Automated Parallelization System and Infrastructure for Multiple Ecosystems☆82Updated last year