meta-pytorch / torchforgeLinks
PyTorch-native post-training at scale
☆613Updated this week
Alternatives and similar repositories for torchforge
Users that are interested in torchforge are comparing it to the libraries listed below
Sorting:
- Miles is an enterprise-facing reinforcement learning framework for LLM and VLM post-training, forked from and co-evolving with slime.☆830Updated this week
- Checkpoint-engine is a simple middleware to update model weights in LLM inference engines☆902Updated last week
- Training API and CLI☆325Updated last week
- Scalable toolkit for efficient model reinforcement☆1,293Updated this week
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.☆334Updated 3 months ago
- ☆957Updated 3 months ago
- Pytorch Distributed native training library for LLMs/VLMs with OOTB Hugging Face support☆266Updated this week
- A scalable asynchronous reinforcement learning implementation with in-flight weight updates.☆361Updated this week
- Load compute kernels from the Hub☆397Updated this week
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)☆475Updated this week
- Simple & Scalable Pretraining for Neural Architecture Research☆307Updated 2 months ago
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)☆273Updated this week
- ☆232Updated 2 months ago
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆279Updated 2 months ago
- Our first fully AI generated deep learning system☆481Updated last week
- Dion optimizer algorithm☆424Updated 3 weeks ago
- TPU inference for vLLM, with unified JAX and PyTorch support.☆228Updated this week
- Async RL Training at Scale☆1,044Updated this week
- A project to improve skills of large language models☆813Updated this week
- ☆219Updated last year
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆371Updated last year
- JAX backend for SGL☆234Updated this week
- Accelerating MoE with IO and Tile-aware Optimizations☆569Updated 3 weeks ago
- An early research stage expert-parallel load balancer for MoE models based on linear programming.☆495Updated 2 months ago
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark + Toolkit with Torch -> CUDA (+ more DSLs)☆792Updated 2 weeks ago
- LLM KV cache compression made easy☆876Updated last week
- Block Diffusion for Ultra-Fast Speculative Decoding☆459Updated this week
- Training library for Megatron-based models with bidirectional Hugging Face conversion capability☆419Updated this week
- Memory optimized Mixture of Experts☆73Updated 6 months ago
- ☆579Updated 4 months ago