NVIDIA-NeMo / SkillsLinks
A project to improve skills of large language models
β619Updated last week
Alternatives and similar repositories for Skills
Users that are interested in Skills are comparing it to the libraries listed below
Sorting:
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β573Updated last month
- Scalable toolkit for efficient model alignmentβ847Updated last month
- PyTorch building blocks for the OLMo ecosystemβ400Updated last week
- Reproducible, flexible LLM evaluationsβ286Updated last week
- β1,010Updated 4 months ago
- Automatic evals for LLMsβ558Updated 5 months ago
- OLMoE: Open Mixture-of-Experts Language Modelsβ916Updated 2 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β358Updated 11 months ago
- Recipes to scale inference-time compute of open modelsβ1,118Updated 6 months ago
- [ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"β441Updated last year
- SkyRL: A Modular Full-stack RL Library for LLMsβ1,287Updated this week
- Tina: Tiny Reasoning Models via LoRAβ308Updated 2 months ago
- Scalable toolkit for efficient model reinforcementβ1,048Updated this week
- Official repository for ORPOβ467Updated last year
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ456Updated 6 months ago
- β555Updated last year
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β750Updated last year
- β474Updated last year
- β327Updated 6 months ago
- RewardBench: the first evaluation tool for reward models.β660Updated 5 months ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuningβ364Updated last year
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β623Updated 8 months ago
- Implementation of paper Data Engineering for Scaling Language Models to 128K Contextβ478Updated last year
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"β268Updated last month
- Large Reasoning Modelsβ807Updated 11 months ago
- β912Updated 3 weeks ago
- β966Updated 10 months ago
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data β¦β793Updated 8 months ago
- LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.β272Updated last month
- A simple unified framework for evaluating LLMsβ254Updated 7 months ago