NVIDIA-NeMo / SkillsLinks
A project to improve skills of large language models
β665Updated this week
Alternatives and similar repositories for Skills
Users that are interested in Skills are comparing it to the libraries listed below
Sorting:
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β582Updated last month
- Scalable toolkit for efficient model alignmentβ846Updated 2 months ago
- PyTorch building blocks for the OLMo ecosystemβ563Updated this week
- Reproducible, flexible LLM evaluationsβ305Updated last month
- Automatic evals for LLMsβ567Updated 5 months ago
- β1,035Updated 5 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β360Updated last year
- Scalable toolkit for efficient model reinforcementβ1,141Updated this week
- Tina: Tiny Reasoning Models via LoRAβ310Updated 2 months ago
- OLMoE: Open Mixture-of-Experts Language Modelsβ930Updated 2 months ago
- Recipes to scale inference-time compute of open modelsβ1,120Updated 7 months ago
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ463Updated 7 months ago
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β631Updated 9 months ago
- Official repository for ORPOβ468Updated last year
- SkyRL: A Modular Full-stack RL Library for LLMsβ1,394Updated this week
- β558Updated last year
- β477Updated last year
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).β332Updated this week
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β751Updated last year
- β937Updated last month
- Large Reasoning Modelsβ806Updated last year
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"β270Updated 2 months ago
- [ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"β446Updated last year
- [ICML 2025 Oral] CodeI/O: Condensing Reasoning Patterns via Code Input-Output Predictionβ564Updated 7 months ago
- An Open Source Toolkit For LLM Distillationβ810Updated this week
- Official repository for the paper "LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code"β745Updated 5 months ago
- [NeurIPS 2025 Spotlight] Reasoning Environments for Reinforcement Learning with Verifiable Rewardsβ1,276Updated last week
- β969Updated 10 months ago
- Minimal hackable GRPO implementationβ306Updated 10 months ago
- Code for Paper: Training Software Engineering Agents and Verifiers with SWE-Gym [ICML 2025]β601Updated 4 months ago