Lightning-AI / forked-pdbLinks
Python pdb for multiple processes
☆79Updated 8 months ago
Alternatives and similar repositories for forked-pdb
Users that are interested in forked-pdb are comparing it to the libraries listed below
Sorting:
- Odysseus: Playground of LLM Sequence Parallelism☆79Updated last year
- Triton implementation of FlashAttention2 that adds Custom Masks.☆163Updated last year
- pytorch-profiler☆50Updated 2 years ago
- ☆160Updated 2 years ago
- Best practices for training DeepSeek, Mixtral, Qwen and other MoE models using Megatron Core.☆158Updated last week
- PyTorch bindings for CUTLASS grouped GEMM.☆141Updated 8 months ago
- Fairring (FAIR + Herring) is a plug-in for PyTorch that provides a process group for distributed training that outperforms NCCL at large …☆65Updated 3 years ago
- ☆115Updated last year
- Official repository for DistFlashAttn: Distributed Memory-efficient Attention for Long-context LLMs Training☆222Updated last year
- PyTorch bindings for CUTLASS grouped GEMM.☆185Updated last month
- Repository of the paper "Accelerating Transformer Inference for Translation via Parallel Decoding"☆123Updated last year
- ☆45Updated 2 years ago
- Flash-Muon: An Efficient Implementation of Muon Optimizer☆229Updated 7 months ago
- Examples for MS-AMP package.☆30Updated 6 months ago
- 🚀 Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.☆219Updated this week
- ring-attention experiments☆165Updated last year
- Megatron's multi-modal data loader☆310Updated this week
- [ICLR 2025] COAT: Compressing Optimizer States and Activation for Memory-Efficient FP8 Training☆258Updated 5 months ago
- This repository contains the experimental PyTorch native float8 training UX☆227Updated last year
- ☆132Updated 8 months ago
- Accelerate LLM preference tuning via prefix sharing with a single line of code☆51Updated 6 months ago
- ☆117Updated 8 months ago
- Triton-based implementation of Sparse Mixture of Experts.☆263Updated 3 months ago
- ☆124Updated last year
- Tritonbench is a collection of PyTorch custom operators with example inputs to measure their performance.☆319Updated this week
- Training neural networks in TensorFlow 2.0 with 5x less memory☆137Updated 3 years ago
- Torch Distributed Experimental☆117Updated last year
- A collection of memory efficient attention operators implemented in the Triton language.☆287Updated last year
- Autonomous GPU Kernel Generation via Deep Agents☆223Updated this week
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mind…☆164Updated 3 weeks ago