π Pytorch Distributed native training library for LLMs/VLMs with OOTB Hugging Face support
β454Apr 25, 2026Updated this week
Alternatives and similar repositories for Automodel
Users that are interested in Automodel are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Training library for Megatron-based models with bidirectional Hugging Face conversion capabilityβ599Updated this week
- A library for exporting models including NeMo and Hugging Face to optimized inference backends, and deploying them for efficient queryingβ33Apr 23, 2026Updated last week
- Scalable toolkit for efficient model reinforcementβ1,568Updated this week
- Implementation from scratch in C of the Multi-head latent attention used in the Deepseek-v3 technical paper.β18Jan 15, 2025Updated last year
- Minimalistic large language model 3D-parallelism trainingβ2,663Apr 7, 2026Updated 3 weeks ago
- Deploy to Railway using AI coding agents - Free Credits Offer β’ AdUse Claude Code, Codex, OpenCode, and more. Autonomous software development now has the infrastructure to match with Railway.
- The tool facilitates debugging convergence issues and testing new algorithms and recipes for training LLMs using Nvidia libraries such asβ¦β19Sep 17, 2025Updated 7 months ago
- [Archived] For the latest updates and community contribution, please visit: https://github.com/Ascend/TransferQueue or https://gitcode.coβ¦β15Jan 16, 2026Updated 3 months ago
- Efficient Long-context Language Model Training by Core Attention Disaggregationβ98Apr 7, 2026Updated 3 weeks ago
- Bridge Megatron-Core to Hugging Face/Reinforcement Learningβ210Updated this week
- A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit and 4-bit floating point (FP8 and FP4) precision on Hβ¦β3,291Updated this week
- Ongoing research training transformer models at scaleβ18Apr 9, 2026Updated 3 weeks ago
- Scalable data pre processing and curation toolkit for LLMsβ1,538Apr 23, 2026Updated last week
- GPU-optimized framework for training diffusion language models at any scale. The backend of Quokka, Super Data Learners, and OpenMoE 2 trβ¦β331Nov 11, 2025Updated 5 months ago
- Accelerating MoE with IO and Tile-aware Optimizationsβ661Apr 22, 2026Updated last week
- GPU virtual machines on DigitalOcean Gradient AI β’ AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- Byted PyTorch Distributed for Hyperscale Training of LLMs and RLsβ1,009Mar 3, 2026Updated last month
- End-to-end pipeline for PPIFlowβ24Feb 17, 2026Updated 2 months ago
- The official codebase for "Experiential Reinforcement Learning" - https://arxiv.org/pdf/2602.13949v1β68Apr 8, 2026Updated 3 weeks ago
- β49May 20, 2025Updated 11 months ago
- A PyTorch native platform for training generative AI modelsβ5,258Apr 23, 2026Updated last week
- β33Apr 19, 2025Updated last year
- Tiny-FSDP, a minimalistic re-implementation of the PyTorch FSDPβ105Aug 20, 2025Updated 8 months ago
- NVIDIA NVSHMEM is a parallel programming interface for NVIDIA GPUs based on OpenSHMEM. NVSHMEM can significantly reduce multi-process comβ¦β515Apr 14, 2026Updated 2 weeks ago
- Miles is an enterprise-facing reinforcement learning framework for LLM and VLM post-training, forked from and co-evolving with slime.β1,145Updated this week
- AI Agents on DigitalOcean Gradient AI Platform β’ AdBuild production-ready AI agents using customizable tools or access multiple LLMs through a single endpoint. Create custom knowledge bases or connect external data.
- [NAACL'25 π SAC Award] Official code for "Advancing MoE Efficiency: A Collaboration-Constrained Routing (C2R) Strategy for Better Expertβ¦β16Feb 4, 2025Updated last year
- a simple API to use CUPTIβ10Aug 19, 2025Updated 8 months ago
- β32Dec 31, 2025Updated 4 months ago
- Utility scripts for PyTorch (e.g. Make Perfetto show some disappearing kernels, Memory profiler that understands more low-level allocatioβ¦β105Sep 11, 2025Updated 7 months ago
- A high-performance RL training-inference weight synchronization framework, designed to enable second-level parameter updates from traininβ¦β150Apr 22, 2026Updated last week
- β12Apr 16, 2026Updated 2 weeks ago
- Ongoing research training transformer models at scaleβ16,145Updated this week
- A tool to configure, launch and manage your machine learning experiments.β241Updated this week
- DLSlime: Flexible & Efficient Heterogeneous Transfer Toolkitβ98Apr 23, 2026Updated last week
- Deploy to Railway using AI coding agents - Free Credits Offer β’ AdUse Claude Code, Codex, OpenCode, and more. Autonomous software development now has the infrastructure to match with Railway.
- A lightweight, user-friendly data-plane for LLM training.β39Sep 10, 2025Updated 7 months ago
- DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scalingβ22Updated this week
- A lightweight, AI-native training framework for large language models. Designed for fast iteration, reproducible experiments, and modularβ¦β559Updated this week
- β180Updated this week
- Ship correct and fast LLM kernels to PyTorchβ148Jan 14, 2026Updated 3 months ago
- Best practices for training DeepSeek, Mixtral, Qwen and other MoE models using Megatron Core.β186Mar 17, 2026Updated last month
- Large Context Attentionβ770Oct 13, 2025Updated 6 months ago