NVIDIA-NeMo / SkillsLinks
A project to improve skills of large language models
β813Updated this week
Alternatives and similar repositories for Skills
Users that are interested in Skills are comparing it to the libraries listed below
Sorting:
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β627Updated last week
- Scalable toolkit for efficient model reinforcementβ1,293Updated this week
- Scalable toolkit for efficient model alignmentβ852Updated 4 months ago
- PyTorch building blocks for the OLMo ecosystemβ785Updated this week
- OLMoE: Open Mixture-of-Experts Language Modelsβ965Updated 4 months ago
- SkyRL: A Modular Full-stack RL Library for LLMsβ1,547Updated this week
- β1,088Updated last month
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β371Updated last year
- Reproducible, flexible LLM evaluationsβ337Updated 2 weeks ago
- Automatic evals for LLMsβ579Updated last month
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β675Updated 10 months ago
- [ICLR 2026] Tina: Tiny Reasoning Models via LoRAβ319Updated 4 months ago
- β1,388Updated 4 months ago
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ468Updated 8 months ago
- Large Reasoning Modelsβ807Updated last year
- β961Updated 3 months ago
- A family of compressed models obtained via pruning and knowledge distillationβ364Updated 3 months ago
- β481Updated last year
- Official repository for the paper "LiveCodeBench: Holistic and Contamination Free Evaluation of Large Language Models for Code"β786Updated 6 months ago
- [ICML 2025 Oral] CodeI/O: Condensing Reasoning Patterns via Code Input-Output Predictionβ567Updated 9 months ago
- Recipes to scale inference-time compute of open modelsβ1,124Updated 8 months ago
- [ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"β445Updated last year
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"β273Updated 3 months ago
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).β344Updated last month
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data β¦β826Updated 10 months ago
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β752Updated last year
- β330Updated 8 months ago
- LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.β278Updated 3 months ago
- Understanding R1-Zero-Like Training: A Critical Perspectiveβ1,205Updated 5 months ago
- β564Updated last year