DS3Lab / DT-FM
☆91Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for DT-FM
- PyTorch implementation of paper "Response Length Perception and Sequence Scheduling: An LLM-Empowered LLM Inference Pipeline".☆74Updated last year
- A resilient distributed training framework☆85Updated 7 months ago
- ☆70Updated 2 years ago
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆34Updated 2 years ago
- ☆65Updated 3 years ago
- ☆46Updated 5 months ago
- Python package for rematerialization-aware gradient checkpointing☆23Updated last year
- PyTorch library for cost-effective, fast and easy serving of MoE models.☆103Updated 3 months ago
- ☆35Updated 3 months ago
- Official repository for LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers☆195Updated 3 months ago
- ☆19Updated last year
- A ChatGPT(GPT-3.5) & GPT-4 Workload Trace to Optimize LLM Serving Systems☆132Updated last month
- ☆88Updated 2 months ago
- AlpaServe: Statistical Multiplexing with Model Parallelism for Deep Learning Serving (OSDI 23)☆78Updated last year
- NAACL '24 (Best Demo Paper RunnerUp) / MlSys @ NeurIPS '23 - RedCoast: A Lightweight Tool to Automate Distributed Training and Inference☆61Updated 3 weeks ago
- SpotServe: Serving Generative Large Language Models on Preemptible Instances☆101Updated 8 months ago
- Official Repo for "LLM-PQ: Serving LLM on Heterogeneous Clusters with Phase-Aware Partition and Adaptive Quantization"☆27Updated 8 months ago
- Code associated with the paper **Fine-tuning Language Models over Slow Networks using Activation Compression with Guarantees**.☆26Updated last year
- GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM☆147Updated 4 months ago
- nnScaler: Compiling DNN models for Parallel Training☆74Updated 3 weeks ago
- A Cluster-Wide Model Manager to Accelerate DNN Training via Automated Training Warmup☆34Updated last year
- Research and development for optimizing transformers☆125Updated 3 years ago
- PipeTransformer: Automated Elastic Pipelining for Distributed Training of Large-scale Models. ICML 2021☆55Updated 3 years ago
- Flash-LLM: Enabling Cost-Effective and Highly-Efficient Large Generative Model Inference with Unstructured Sparsity☆180Updated last year
- [NeurIPS'23] Speculative Decoding with Big Little Decoder☆86Updated 9 months ago
- ☆51Updated last month
- Triton-based implementation of Sparse Mixture of Experts.☆185Updated last month
- [ICML 2024] Serving LLMs on heterogeneous decentralized clusters.☆15Updated 6 months ago
- Automated Parallelization System and Infrastructure for Multiple Ecosystems☆75Updated this week
- Chimera: Efficiently Training Large-Scale Neural Networks with Bidirectional Pipelines.☆46Updated 11 months ago