NVIDIA-NeMo / SkillsLinks
A project to improve skills of large language models
β786Updated last week
Alternatives and similar repositories for Skills
Users that are interested in Skills are comparing it to the libraries listed below
Sorting:
- PyTorch building blocks for the OLMo ecosystemβ741Updated last week
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β621Updated 3 weeks ago
- Scalable toolkit for efficient model reinforcementβ1,267Updated this week
- Scalable toolkit for efficient model alignmentβ848Updated 3 months ago
- OLMoE: Open Mixture-of-Experts Language Modelsβ961Updated 4 months ago
- Reproducible, flexible LLM evaluationsβ331Updated this week
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β370Updated last year
- Automatic evals for LLMsβ575Updated last month
- β1,080Updated 3 weeks ago
- SkyRL: A Modular Full-stack RL Library for LLMsβ1,518Updated this week
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β667Updated 10 months ago
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β750Updated last year
- Tina: Tiny Reasoning Models via LoRAβ313Updated 4 months ago
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ469Updated 8 months ago
- Large Reasoning Modelsβ807Updated last year
- Recipes to scale inference-time compute of open modelsβ1,124Updated 8 months ago
- A family of compressed models obtained via pruning and knowledge distillationβ364Updated 2 months ago
- β559Updated last year
- β952Updated 2 months ago
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data β¦β825Updated 10 months ago
- β1,383Updated 4 months ago
- LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.β276Updated 3 months ago
- [ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"β445Updated last year
- Official repository for ORPOβ469Updated last year
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).β344Updated last month
- Understanding R1-Zero-Like Training: A Critical Perspectiveβ1,199Updated 5 months ago
- β328Updated 8 months ago
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"β273Updated 3 months ago
- An Open Source Toolkit For LLM Distillationβ846Updated last month
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024β355Updated last week