NVIDIA-NeMo / RunLinks
A tool to configure, launch and manage your machine learning experiments.
β174Updated this week
Alternatives and similar repositories for Run
Users that are interested in Run are comparing it to the libraries listed below
Sorting:
- π Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flashβ¦β258Updated last week
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)β372Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMsβ265Updated 9 months ago
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)β190Updated this week
- Load compute kernels from the Hubβ220Updated this week
- Google TPU optimizations for transformers modelsβ117Updated 6 months ago
- Scalable and Performant Data Loadingβ290Updated last week
- Easy and lightning fast training of π€ Transformers on Habana Gaudi processor (HPU)β191Updated this week
- π Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.β206Updated this week
- ArcticInference: vLLM plugin for high-throughput, low-latency inferenceβ198Updated this week
- β215Updated 6 months ago
- Scalable toolkit for efficient model reinforcementβ558Updated this week
- Accelerating your LLM training to full speed! Made with β€οΈ by ServiceNow Researchβ217Updated this week
- A family of compressed models obtained via pruning and knowledge distillationβ347Updated 8 months ago
- This repository contains the experimental PyTorch native float8 training UXβ224Updated last year
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024β323Updated 3 months ago
- β280Updated this week
- Efficient LLM Inference over Long Sequencesβ385Updated last month
- A safetensors extension to efficiently store sparse quantized tensors on diskβ141Updated this week
- ποΈ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of Oβ¦β307Updated 2 months ago
- PyTorch Single Controllerβ341Updated this week
- Megatron's multi-modal data loaderβ230Updated last week
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.β198Updated last year
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"β66Updated 4 months ago
- Easy and Efficient Quantization for Transformersβ198Updated last month
- LLM KV cache compression made easyβ560Updated last week
- β206Updated 5 months ago
- β114Updated last year
- β120Updated last year
- Applied AI experiments and examples for PyTorchβ289Updated 2 months ago