NVIDIA-NeMo / SkillsLinks
A project to improve skills of large language models
β608Updated this week
Alternatives and similar repositories for Skills
Users that are interested in Skills are comparing it to the libraries listed below
Sorting:
- Scalable toolkit for efficient model alignmentβ844Updated last month
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β558Updated last week
- PyTorch building blocks for the OLMo ecosystemβ317Updated this week
- β995Updated 4 months ago
- Scalable toolkit for efficient model reinforcementβ1,009Updated this week
- Reproducible, flexible LLM evaluationsβ264Updated 2 weeks ago
- Automatic evals for LLMsβ556Updated 4 months ago
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β353Updated 11 months ago
- Official repository for ORPOβ463Updated last year
- SkyRL: A Modular Full-stack RL Library for LLMsβ1,170Updated this week
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β750Updated last year
- OLMoE: Open Mixture-of-Experts Language Modelsβ899Updated last month
- Recipes to scale inference-time compute of open modelsβ1,117Updated 5 months ago
- β552Updated 11 months ago
- Tina: Tiny Reasoning Models via LoRAβ304Updated last month
- [ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"β441Updated last year
- An extension of the nanoGPT repository for training small MOE models.β210Updated 8 months ago
- RewardBench: the first evaluation tool for reward models.β649Updated 5 months ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuningβ365Updated last year
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"β269Updated 3 weeks ago
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data β¦β786Updated 7 months ago
- Large Reasoning Modelsβ806Updated 11 months ago
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ450Updated 5 months ago
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β616Updated 7 months ago
- Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.β443Updated last year
- Implementation of paper Data Engineering for Scaling Language Models to 128K Contextβ477Updated last year
- A simplified implementation for experimenting with RLVR on GSM8K, This repository provides a starting point for exploring reasoning.β144Updated 9 months ago
- LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.β264Updated 2 weeks ago
- An Open Source Toolkit For LLM Distillationβ777Updated 4 months ago
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).β296Updated 2 weeks ago