NVIDIA-NeMo / SkillsLinks
A project to improve skills of large language models
β581Updated last week
Alternatives and similar repositories for Skills
Users that are interested in Skills are comparing it to the libraries listed below
Sorting:
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β535Updated this week
- PyTorch building blocks for the OLMo ecosystemβ305Updated last week
- Scalable toolkit for efficient model alignmentβ842Updated last week
- Reproducible, flexible LLM evaluationsβ256Updated this week
- Automatic evals for LLMsβ543Updated 3 months ago
- β963Updated 3 months ago
- Scalable toolkit for efficient model reinforcementβ931Updated this week
- SkyRL: A Modular Full-stack RL Library for LLMsβ1,060Updated this week
- OLMoE: Open Mixture-of-Experts Language Modelsβ886Updated 3 weeks ago
- Recipes to scale inference-time compute of open modelsβ1,109Updated 4 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β342Updated 10 months ago
- β543Updated 11 months ago
- Official repository for ORPOβ463Updated last year
- [ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"β440Updated last year
- Tina: Tiny Reasoning Models via LoRAβ296Updated 3 weeks ago
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β747Updated last year
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuningβ362Updated last year
- β321Updated 4 months ago
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β607Updated 7 months ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"β311Updated last year
- β465Updated last year
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ447Updated 5 months ago
- RewardBench: the first evaluation tool for reward models.β642Updated 4 months ago
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data β¦β778Updated 7 months ago
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"β265Updated 5 months ago
- Large Reasoning Modelsβ805Updated 10 months ago
- Implementation of paper Data Engineering for Scaling Language Models to 128K Contextβ477Updated last year
- β827Updated this week
- A simple unified framework for evaluating LLMsβ251Updated 6 months ago
- An extension of the nanoGPT repository for training small MOE models.β197Updated 7 months ago