NVIDIA / NeMo-RLLinks
Scalable toolkit for efficient model reinforcement
β361Updated this week
Alternatives and similar repositories for NeMo-RL
Users that are interested in NeMo-RL are comparing it to the libraries listed below
Sorting:
- π Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flashβ¦β249Updated this week
- β188Updated 3 months ago
- A project to improve skills of large language modelsβ413Updated this week
- Super-Efficient RLHF Training of LLMs with Parameter Reallocationβ299Updated last month
- SkyRL-v0: Train Real-World Long-Horizon Agents via Reinforcement Learningβ343Updated last week
- ByteCheckpoint: An Unified Checkpointing Library for LFMsβ215Updated 2 months ago
- LLM KV cache compression made easyβ493Updated 3 weeks ago
- π₯ A minimal training framework for scaling FLA modelsβ146Updated 3 weeks ago
- Triton-based implementation of Sparse Mixture of Experts.β216Updated 6 months ago
- [ICLR 2025] COAT: Compressing Optimizer States and Activation for Memory-Efficient FP8 Trainingβ203Updated 2 weeks ago
- Ring attention implementation with flash attentionβ771Updated last week
- Scalable toolkit for efficient model alignmentβ803Updated 2 weeks ago
- Triton implementation of FlashAttention2 that adds Custom Masks.β117Updated 9 months ago
- Official repository for DistFlashAttn: Distributed Memory-efficient Attention for Long-context LLMs Trainingβ208Updated 9 months ago
- Efficient triton implementation of Native Sparse Attention.β155Updated last week
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β367Updated this week
- VeOmni: Scaling any Modality Model Training to any Accelerators with PyTorch native Training Frameworkβ339Updated 3 weeks ago
- Async pipelined version of Verlβ91Updated last month
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β333Updated 5 months ago
- A scalable asynchronous reinforcement learning implementation with in-flight weight updates.β117Updated this week
- Large Context Attentionβ711Updated 4 months ago
- kernels, of the mega varietyβ184Updated this week
- Explorations into some recent techniques surrounding speculative decodingβ266Updated 5 months ago
- Normalized Transformer (nGPT)β181Updated 6 months ago
- Efficient LLM Inference over Long Sequencesβ376Updated this week
- Zero Bubble Pipeline Parallelismβ395Updated 3 weeks ago
- REST: Retrieval-Based Speculative Decoding, NAACL 2024β202Updated 6 months ago
- [ICML 2024] Quest: Query-Aware Sparsity for Efficient Long-Context LLM Inferenceβ291Updated 6 months ago
- π Collection of components for development, training, tuning, and inference of foundation models leveraging PyTorch native components.β196Updated this week
- β450Updated this week