Youhe-Jiang / IJCAI2023-OptimalShardedDataParallelView external linksLinks
[IJCAI2023] An automated parallel training system that combines the advantages from both data and model parallelism. If you have any interests, please visit/star/fork https://github.com/Youhe-Jiang/OptimalShardedDataParallel
☆52May 31, 2023Updated 2 years ago
Alternatives and similar repositories for IJCAI2023-OptimalShardedDataParallel
Users that are interested in IJCAI2023-OptimalShardedDataParallel are comparing it to the libraries listed below
Sorting:
- DELTA-pytorch:DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation☆12Apr 16, 2024Updated last year
- An MPI wrapper for the pytorch tensor library that is automatically differentiable☆10Mar 27, 2023Updated 2 years ago
- ☆24Jul 7, 2024Updated last year
- A schedule language for large model training☆152Aug 21, 2025Updated 5 months ago
- ☆84Dec 2, 2022Updated 3 years ago
- An implementation of parameter server framework in PyTorch RPC.☆12Nov 12, 2021Updated 4 years ago
- ☆14Aug 29, 2023Updated 2 years ago
- ☆12Apr 22, 2022Updated 3 years ago
- ☆14Jan 12, 2022Updated 4 years ago
- ☆77May 4, 2021Updated 4 years ago
- ☆23Aug 21, 2025Updated 5 months ago
- Zero Bubble Pipeline Parallelism☆449May 7, 2025Updated 9 months ago
- CFR implementation of a poker bot.☆12Feb 17, 2023Updated 2 years ago
- ☆82Updated this week
- Distributed IO-aware Attention algorithm☆24Sep 24, 2025Updated 4 months ago
- ☆56Jan 25, 2021Updated 5 years ago
- Standalone Flash Attention v2 kernel without libtorch dependency☆114Sep 10, 2024Updated last year
- Implementations of a large collection of reinforcement learning algorithms.☆28Nov 30, 2023Updated 2 years ago
- Yinghan's Code Sample☆365Jul 25, 2022Updated 3 years ago
- LLM training technologies developed by kwai