NVIDIA-NeMo / RunLinks
A tool to configure, launch and manage your machine learning experiments.
β210Updated this week
Alternatives and similar repositories for Run
Users that are interested in Run are comparing it to the libraries listed below
Sorting:
- π Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flashβ¦β275Updated last month
- Load compute kernels from the Hubβ352Updated last week
- Pytorch Distributed native training library for LLMs/VLMs with OOTB Hugging Face supportβ214Updated this week
- Google TPU optimizations for transformers modelsβ131Updated last week
- Fault tolerance for PyTorch (HSDP, LocalSGD, DiLoCo, Streaming DiLoCo)β459Updated 2 weeks ago
- Scalable and Performant Data Loadingβ356Updated this week
- β219Updated 11 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMsβ267Updated 3 weeks ago
- ποΈ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of Oβ¦β323Updated 3 months ago
- Easy and lightning fast training of π€ Transformers on Habana Gaudi processor (HPU)β202Updated last week
- π Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.β216Updated 2 weeks ago
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)β261Updated last week
- Accelerating your LLM training to full speed! Made with β€οΈ by ServiceNow Researchβ272Updated last week
- A family of compressed models obtained via pruning and knowledge distillationβ361Updated last month
- This repository contains the experimental PyTorch native float8 training UXβ227Updated last year
- PyTorch/XLA integration with JetStream (https://github.com/google/JetStream) for LLM inference"β79Updated last week
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024β349Updated 7 months ago
- Simple & Scalable Pretraining for Neural Architecture Researchβ305Updated 2 weeks ago
- ArcticInference: vLLM plugin for high-throughput, low-latency inferenceβ354Updated this week
- Efficient LLM Inference over Long Sequencesβ394Updated 6 months ago
- π· Build compute kernelsβ195Updated this week
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.β327Updated last month
- LM engine is a library for pretraining/finetuning LLMsβ102Updated this week
- β321Updated last week
- β205Updated last year
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.β202Updated last year
- β225Updated last month
- Megatron's multi-modal data loaderβ292Updated last week
- β565Updated 3 months ago
- Provides end-to-end model development pipelines for LLMs and Multimodal models that can be launched on-prem or cloud-native.β509Updated 8 months ago