NVIDIA-NeMo / SkillsLinks
A project to improve skills of large language models
β734Updated last week
Alternatives and similar repositories for Skills
Users that are interested in Skills are comparing it to the libraries listed below
Sorting:
- πΎ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.β609Updated this week
- PyTorch building blocks for the OLMo ecosystemβ656Updated this week
- Scalable toolkit for efficient model alignmentβ848Updated 3 months ago
- Scalable toolkit for efficient model reinforcementβ1,210Updated this week
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsβ¦β366Updated last year
- β1,052Updated 6 months ago
- OLMoE: Open Mixture-of-Experts Language Modelsβ945Updated 3 months ago
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β751Updated last year
- Automatic evals for LLMsβ570Updated 2 weeks ago
- Reproducible, flexible LLM evaluationsβ316Updated last month
- [ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"β447Updated last year
- SkyRL: A Modular Full-stack RL Library for LLMsβ1,437Updated this week
- Recipes to scale inference-time compute of open modelsβ1,123Updated 7 months ago
- Official repository for ORPOβ468Updated last year
- [NeurIPS'25] Official codebase for "SWE-RL: Advancing LLM Reasoning via Reinforcement Learning on Open Software Evolution"β650Updated 9 months ago
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ467Updated 7 months ago
- Tina: Tiny Reasoning Models via LoRAβ312Updated 3 months ago
- A family of compressed models obtained via pruning and knowledge distillationβ362Updated 2 months ago
- β559Updated last year
- β945Updated 2 months ago
- Large Reasoning Modelsβ804Updated last year
- β478Updated last year
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).β340Updated 3 weeks ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuningβ366Updated last year
- An Open Source Toolkit For LLM Distillationβ819Updated 3 weeks ago
- An extension of the nanoGPT repository for training small MOE models.β224Updated 10 months ago
- Implementation of paper Data Engineering for Scaling Language Models to 128K Contextβ482Updated last year
- Single File, Single GPU, From Scratch, Efficient, Full Parameter Tuning library for "RL for LLMs"β575Updated 3 months ago
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"β271Updated 2 months ago
- LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.β276Updated 2 months ago