ROCm / Megatron-LMLinks
Ongoing research training transformer models at scale
☆24Updated this week
Alternatives and similar repositories for Megatron-LM
Users that are interested in Megatron-LM are comparing it to the libraries listed below
Sorting:
- Microsoft Collective Communication Library☆65Updated 8 months ago
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆61Updated last month
- ☆41Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆86Updated last week
- ☆85Updated 9 months ago
- Fast and memory-efficient exact attention☆179Updated last week
- MAD (Model Automation and Dashboarding)☆23Updated last week
- RCCL Performance Benchmark Tests☆71Updated last week
- Development repository for the Triton language and compiler☆127Updated this week
- A CUTLASS implementation using SYCL☆32Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆78Updated this week
- oneCCL Bindings for Pytorch*☆99Updated this week
- extensible collectives library in triton☆88Updated 4 months ago
- Estimate MFU for DeepSeekV3☆25Updated 7 months ago
- AI Tensor Engine for ROCm☆243Updated this week
- An efficient GPU support for LLM inference with x-bit quantization (e.g. FP6,FP5).☆260Updated 3 weeks ago
- ☆67Updated last year
- ☆102Updated 7 months ago
- nnScaler: Compiling DNN models for Parallel Training☆114Updated this week
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆40Updated 2 years ago
- Aims to implement dual-port and multi-qp solutions in deepEP ibrc transport☆58Updated 3 months ago
- PyTorch bindings for CUTLASS grouped GEMM.☆107Updated 2 months ago
- Official repository for DistFlashAttn: Distributed Memory-efficient Attention for Long-context LLMs Training☆212Updated 11 months ago
- LLM-Inference-Bench☆48Updated 3 weeks ago
- Applied AI experiments and examples for PyTorch☆289Updated 2 months ago
- ☆20Updated last week
- A lightweight design for computation-communication overlap.☆155Updated last month
- QuickReduce is a performant all-reduce library designed for AMD ROCm that supports inline compression.☆31Updated 4 months ago
- [ICLR2025] Breaking Throughput-Latency Trade-off for Long Sequences with Speculative Decoding☆123Updated 8 months ago
- This repository contains the results and code for the MLPerf™ Training v2.0 benchmark.☆29Updated last year