Shigangli / Chimera
Chimera: Efficiently Training Large-Scale Neural Networks with Bidirectional Pipelines.
☆46Updated 11 months ago
Related projects ⓘ
Alternatives and complementary repositories for Chimera
- ☆73Updated last year
- ☆65Updated 3 years ago
- AlpaServe: Statistical Multiplexing with Model Parallelism for Deep Learning Serving (OSDI 23)☆78Updated last year
- nnScaler: Compiling DNN models for Parallel Training☆74Updated 3 weeks ago
- An experimental parallel training platform☆52Updated 7 months ago
- ☆9Updated 2 years ago
- A ChatGPT(GPT-3.5) & GPT-4 Workload Trace to Optimize LLM Serving Systems☆132Updated last month
- Compiler for Dynamic Neural Networks☆43Updated last year
- Magicube is a high-performance library for quantized sparse matrix operations (SpMM and SDDMM) of deep learning on Tensor Cores.☆81Updated last year
- ☆46Updated 5 months ago
- ☆74Updated last month
- Automated Parallelization System and Infrastructure for Multiple Ecosystems☆75Updated this week
- LLM serving cluster simulator☆81Updated 6 months ago
- Artifact for OSDI'23: MGG: Accelerating Graph Neural Networks with Fine-grained intra-kernel Communication-Computation Pipelining on Mult…☆37Updated 8 months ago
- ☆52Updated last week
- A resilient distributed training framework☆85Updated 7 months ago
- Synthesizer for optimal collective communication algorithms☆99Updated 7 months ago
- ☆14Updated 5 months ago
- MAGIS: Memory Optimization via Coordinated Graph Transformation and Scheduling for DNN (ASPLOS'24)☆43Updated 5 months ago
- Artifact for PPoPP22 QGTC: Accelerating Quantized GNN via GPU Tensor Core.☆27Updated 2 years ago
- ☆23Updated last year
- ☆19Updated 4 months ago
- Artifacts for our ASPLOS'23 paper ElasticFlow☆52Updated 6 months ago
- Artifact of OSDI '24 paper, ”Llumnix: Dynamic Scheduling for Large Language Model Serving“☆57Updated 5 months ago
- An Efficient Pipelined Data Parallel Approach for Training Large Model☆70Updated 3 years ago
- ☆30Updated 4 months ago
- ☆44Updated 5 years ago
- ☆23Updated 2 years ago
- Paella: Low-latency Model Serving with Virtualized GPU Scheduling☆57Updated 6 months ago
- ☆56Updated 2 years ago