Distributed-AI / PipeTransformerLinks
PipeTransformer: Automated Elastic Pipelining for Distributed Training of Large-scale Models. ICML 2021
☆56Updated 3 years ago
Alternatives and similar repositories for PipeTransformer
Users that are interested in PipeTransformer are comparing it to the libraries listed below
Sorting:
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆38Updated 2 years ago
- [IJCAI2023] An automated parallel training system that combines the advantages from both data and model parallelism. If you have any inte…☆51Updated 2 years ago
- ☆43Updated last year
- Fairring (FAIR + Herring) is a plug-in for PyTorch that provides a process group for distributed training that outperforms NCCL at large …☆65Updated 3 years ago
- ☆73Updated 4 years ago
- ☆83Updated 3 years ago
- ☆42Updated 2 years ago
- Research and development for optimizing transformers☆126Updated 4 years ago
- FTPipe and related pipeline model parallelism research.☆41Updated 2 years ago
- An Efficient Pipelined Data Parallel Approach for Training Large Model☆76Updated 4 years ago
- ddl-benchmarks: Benchmarks for Distributed Deep Learning☆37Updated 5 years ago
- Machine Learning System☆14Updated 5 years ago
- [ICDCS 2023] DeAR: Accelerating Distributed Deep Learning with Fine-Grained All-Reduce Pipelining☆11Updated last year
- pytorch-profiler☆51Updated 2 years ago
- A Deep Learning Cluster Scheduler☆37Updated 4 years ago
- A Cluster-Wide Model Manager to Accelerate DNN Training via Automated Training Warmup☆34Updated 2 years ago
- Chimera: bidirectional pipeline parallelism for efficiently training large-scale models.☆63Updated 2 months ago
- PyTorch implementation of paper "Response Length Perception and Sequence Scheduling: An LLM-Empowered LLM Inference Pipeline".☆86Updated 2 years ago
- [MLSys 2021] IOS: Inter-Operator Scheduler for CNN Acceleration☆200Updated 3 years ago
- Multi-Instance-GPU profiling tool☆58Updated 2 years ago
- Distributed ML Training Benchmarks☆27Updated 2 years ago
- Code associated with the paper **Fine-tuning Language Models over Slow Networks using Activation Compression with Guarantees**.☆28Updated 2 years ago
- Training neural networks in TensorFlow 2.0 with 5x less memory☆131Updated 3 years ago
- SparseTIR: Sparse Tensor Compiler for Deep Learning☆138Updated 2 years ago
- nnScaler: Compiling DNN models for Parallel Training☆113Updated last month
- ☆146Updated 10 months ago
- sensAI: ConvNets Decomposition via Class Parallelism for Fast Inference on Live Data☆64Updated 10 months ago
- Practical low-rank gradient compression for distributed optimization: https://arxiv.org/abs/1905.13727☆146Updated 7 months ago
- Memory Optimizations for Deep Learning (ICML 2023)☆64Updated last year
- Odysseus: Playground of LLM Sequence Parallelism☆69Updated 11 months ago