NVIDIA-NeMo / RLLinks
Scalable toolkit for efficient model reinforcement
☆626Updated last week
Alternatives and similar repositories for RL
Users that are interested in RL are comparing it to the libraries listed below
Sorting:
- Scalable toolkit for efficient model alignment☆837Updated 3 weeks ago
- A project to improve skills of large language models☆529Updated this week
- SkyRL: A Modular Full-stack RL Library for LLMs☆738Updated this week
- 🚀 Efficiently (pre)training foundation models with native PyTorch features, including FSDP for training and SDPA implementation of Flash…☆260Updated last month
- ☆211Updated 6 months ago
- Ring attention implementation with flash attention☆841Updated 3 weeks ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, spars…☆344Updated 8 months ago
- ☆514Updated 3 weeks ago
- Super-Efficient RLHF Training of LLMs with Parameter Reallocation☆307Updated 4 months ago
- Large Context Attention☆727Updated 7 months ago
- 🌾 OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.☆433Updated last week
- Efficient LLM Inference over Long Sequences☆389Updated 2 months ago
- LLM KV cache compression made easy☆586Updated this week
- slime is a LLM post-training framework aiming for RL Scaling.☆1,375Updated last week
- 🐳 Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"☆822Updated 5 months ago
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems☆527Updated this week
- Decentralized RL Training at Scale☆441Updated this week
- An extension of the nanoGPT repository for training small MOE models.☆178Updated 5 months ago
- PyTorch building blocks for the OLMo ecosystem☆274Updated this week
- OLMoE: Open Mixture-of-Experts Language Models☆842Updated 5 months ago
- TransMLA: Multi-Head Latent Attention Is All You Need☆339Updated last month
- [ICLR 2025] DuoAttention: Efficient Long-Context LLM Inference with Retrieval and Streaming Heads☆487Updated 6 months ago
- Implementation of 💍 Ring Attention, from Liu et al. at Berkeley AI, in Pytorch☆536Updated 3 months ago
- Parallel Scaling Law for Language Model — Beyond Parameter and Inference Time Scaling☆429Updated 3 months ago
- Explorations into some recent techniques surrounding speculative decoding☆282Updated 8 months ago
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)☆200Updated last week
- Single File, Single GPU, From Scratch, Efficient, Full Parameter Tuning library for "RL for LLMs"☆516Updated last month
- ByteCheckpoint: An Unified Checkpointing Library for LFMs☆237Updated last month
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.☆250Updated 2 weeks ago
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆327Updated 3 months ago