NVIDIA / NeMo-SkillsLinks
A project to improve skills of large language models
โ423Updated this week
Alternatives and similar repositories for NeMo-Skills
Users that are interested in NeMo-Skills are comparing it to the libraries listed below
Sorting:
- ๐พ OAT: A research-friendly framework for LLM online alignment, including reinforcement learning, preference learning, etc.โ379Updated last week
- Scalable toolkit for efficient model alignmentโ814Updated 3 weeks ago
- โ773Updated last month
- Scalable toolkit for efficient model reinforcementโ438Updated this week
- Reproducible, flexible LLM evaluationsโ213Updated last month
- โ297Updated 3 weeks ago
- RewardBench: the first evaluation tool for reward models.โ604Updated last week
- [ICML'24] Data and code for our paper "Training-Free Long-Context Scaling of Large Language Models"โ410Updated 8 months ago
- Repo for Rho-1: Token-level Data Selection & Selective Pretraining of LLMs.โ421Updated last year
- Automatic evals for LLMsโ429Updated 2 weeks ago
- Implementation of paper Data Engineering for Scaling Language Models to 128K Contextโ463Updated last year
- PyTorch building blocks for the OLMo ecosystemโ234Updated this week
- Memory layers use a trainable key-value lookup mechanism to add extra parameters to a model without increasing FLOPs. Conceptually, sparsโฆโ337Updated 6 months ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"โ303Updated last year
- SkyRL-v0: Train Real-World Long-Horizon Agents via Reinforcement Learningโ422Updated this week
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.โ731Updated 8 months ago
- A framework to study AI models in Reasoning, Alignment, and use of Memory (RAM).โ255Updated this week
- โ331Updated 2 weeks ago
- A simple toolkit for benchmarking LLMs on mathematical reasoning tasks. ๐งฎโจโ226Updated last year
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuningโ357Updated 9 months ago
- Official repository for ORPOโ455Updated last year
- Official repo for paper: "Reinforcement Learning for Reasoning in Small LLMs: What Works and What Doesn't"โ236Updated last month
- Tina: Tiny Reasoning Models via LoRAโ258Updated 3 weeks ago
- Large Reasoning Modelsโ804Updated 6 months ago
- L1: Controlling How Long A Reasoning Model Thinks With Reinforcement Learningโ220Updated last month
- The official evaluation suite and dynamic data release for MixEval.โ242Updated 7 months ago
- OLMoE: Open Mixture-of-Experts Language Modelsโ785Updated 3 months ago
- LOFT: A 1 Million+ Token Long-Context Benchmarkโ201Updated last week
- LongRoPE is a novel method that can extends the context window of pre-trained LLMs to an impressive 2048k tokens.โ231Updated 9 months ago
- โ288Updated 10 months ago