NVIDIA-NeMo / RunLinks
A tool to configure, launch and manage your machine learning experiments.
β195Updated this week
Alternatives and similar repositories for Run
Users that are interested in Run are comparing it to the libraries listed below
Sorting:
- π Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flashβ¦β268Updated 2 months ago
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)β414Updated this week
- Load compute kernels from the Hubβ290Updated last week
- Scalable and Performant Data Loadingβ304Updated 2 weeks ago
- Google TPU optimizations for transformers modelsβ120Updated 8 months ago
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)β219Updated this week
- DTensor-native pretraining and fine-tuning for LLMs/VLMs with day-0 Hugging Face support, GPU-acceleration, and memory efficiency.β84Updated this week
- β217Updated 8 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMsβ266Updated 11 months ago
- ArcticInference: vLLM plugin for high-throughput, low-latency inferenceβ267Updated this week
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"β73Updated 3 weeks ago
- Easy and lightning fast training of π€ Transformers on Habana Gaudi processor (HPU)β198Updated this week
- Accelerating your LLM training to full speed! Made with β€οΈ by ServiceNow Researchβ237Updated this week
- π Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.β213Updated this week
- π· Build compute kernelsβ149Updated this week
- Simple & Scalable Pretraining for Neural Architecture Researchβ296Updated last month
- This repository contains the experimental PyTorch native float8 training UXβ224Updated last year
- Efficient LLM Inference over Long Sequencesβ391Updated 3 months ago
- ποΈ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of Oβ¦β315Updated last week
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.β280Updated last month
- A family of compressed models obtained via pruning and knowledge distillationβ352Updated 10 months ago
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024β338Updated 5 months ago
- Provides end-to-end model development pipelines for LLMs and Multimodal models that can be launched on-prem or cloud-native.β507Updated 5 months ago
- β221Updated 7 months ago
- β298Updated last week
- β122Updated last year
- Megatron's multi-modal data loaderβ250Updated last week
- PyTorch Single Controllerβ425Updated this week
- PyTorch building blocks for the OLMo ecosystemβ301Updated last week
- An extension of the nanoGPT repository for training small MOE models.β195Updated 6 months ago