NVIDIA-NeMo / SkillsLinks
A project to improve skills of large language models
β813Updated this week
Alternatives and similar repositories for Skills
Users that are interested in Skills are comparing it to the libraries listed below
Sorting:
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β623Updated last week
- Scalable toolkit for efficient model alignmentβ852Updated 4 months ago
- PyTorch building blocks for the OLMo ecosystemβ785Updated this week
- β1,088Updated last month
- Scalable toolkit for efficient model reinforcementβ1,293Updated this week
- Automatic evals for LLMsβ579Updated last month
- Reproducible, flexible LLM evaluationsβ337Updated 2 weeks ago
- [ICLR 2026] Tina: Tiny Reasoning Models via LoRAβ319Updated 4 months ago
- OLMoE: Open Mixture-of-Experts Language Modelsβ965Updated 4 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β371Updated last year
- Large Reasoning Modelsβ807Updated last year
- Recipes to scale inference-time compute of open modelsβ1,124Updated 8 months ago
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ468Updated 8 months ago
- An extension of the nanoGPT repository for training small MOE models.β236Updated 11 months ago
- β1,388Updated 4 months ago
- SkyRL: A Modular Full-stack RL Library for LLMsβ1,547Updated this week
- β970Updated last year
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β751Updated last year
- [ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"β445Updated last year
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β675Updated 10 months ago
- LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.β278Updated 3 months ago
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).β344Updated last month
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"β273Updated 3 months ago
- β481Updated last year
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuningβ366Updated last year
- [ICML 2025 Oral] CodeI/O: Condensing Reasoning Patterns via Code Input-Output Predictionβ567Updated 9 months ago
- [NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewardsβ1,332Updated 3 weeks ago
- β564Updated last year
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data β¦β826Updated 10 months ago
- Official codebase for "Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling".β283Updated 11 months ago