Automatically Discovering Fast Parallelization Strategies for Distributed Deep Neural Network Training
☆1,872Apr 15, 2026Updated this week
Alternatives and similar repositories for flexflow-train
Users that are interested in flexflow-train are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- FlashInfer: Kernel Library for LLM Serving☆5,372Apr 11, 2026Updated last week
- The Tensor Algebra SuperOptimizer for Deep Learning☆741Jan 26, 2023Updated 3 years ago
- Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads☆2,722Jun 25, 2024Updated last year
- Training and serving large-scale neural networks with auto parallelization.☆3,187Dec 9, 2023Updated 2 years ago
- A throughput-oriented high-performance serving framework for LLMs☆952Mar 29, 2026Updated 3 weeks ago
- Bare Metal GPUs on DigitalOcean Gradient AI • AdPurpose-built for serious AI teams training foundational models, running large-scale inference, and pushing the boundaries of what's possible.
- Transformer related optimization, including BERT, GPT☆6,412Mar 27, 2024Updated 2 years ago
- A flexible and efficient deep neural network (DNN) compiler that generates high-performance executable from a DNN model description.☆1,001Sep 19, 2024Updated last year
- MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.☆2,107Jun 30, 2025Updated 9 months ago
- Serving multiple LoRA finetuned LLM as one☆1,152May 8, 2024Updated last year
- Dynamic Memory Management for Serving LLMs without PagedAttention☆478May 30, 2025Updated 10 months ago
- LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalabili…☆4,016Updated this week
- [ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding