☆2,551May 19, 2024Updated last year
Alternatives and similar repositories for weak-to-strong
Users that are interested in weak-to-strong are comparing it to the libraries listed below
Sorting:
- Robust recipes to align language models with human and AI preferences☆5,510Sep 8, 2025Updated 6 months ago
- ☆4,110Jun 4, 2024Updated last year
- 800,000 step-level correctness labels on LLM solutions to MATH problems☆2,096Jun 1, 2023Updated 2 years ago
- Train transformer language models with reinforcement learning.☆17,523Updated this week
- Large World Model -- Modeling Text and Video with Millions Context☆7,399Oct 19, 2024Updated last year
- ☆4,390Jul 31, 2025Updated 7 months ago
- Official inference library for Mistral models☆10,700Feb 26, 2026Updated last week
- Consistency Distilled Diff VAE☆2,211Nov 7, 2023Updated 2 years ago
- Modeling, training, eval, and inference code for OLMo☆6,353Nov 24, 2025Updated 3 months ago
- [ICLR 2024] Efficient Streaming Language Models with Attention Sinks☆7,196Jul 11, 2024Updated last year
- All things prompt engineering☆5,737Jun 4, 2024Updated last year
- Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.☆6,185Aug 22, 2025Updated 6 months ago
- Reference implementation for DPO (Direct Preference Optimization)☆2,861Aug 11, 2024Updated last year
- AllenAI's post-training codebase☆3,614Updated this week
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.☆24,500Aug 12, 2024Updated last year
- An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & TIS & vLLM & Ray & Async RL)☆9,084Updated this week
- Benchmarking large language models' complex reasoning ability with chain-of-thought prompting☆2,766Aug 4, 2024Updated last year
- Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We als…☆18,234Updated this week
- RewardBench: the first evaluation tool for reward models.☆702Feb 16, 2026Updated 3 weeks ago
- A framework for few-shot evaluation of language models.☆11,618Updated this week
- ☆1,074Mar 6, 2024Updated 2 years ago
- verl: Volcano Engine Reinforcement Learning for LLMs☆19,739Updated this week
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆4,738Jan 8, 2024Updated 2 years ago
- Medusa: Simple Framework for Accelerating LLM Generation with Multiple Decoding Heads☆2,714Jun 25, 2024Updated last year
- [ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding☆1,317Mar 6, 2025Updated last year
- Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"☆1,824Jun 17, 2025Updated 8 months ago
- A series of large language models trained from scratch by developers @01-ai☆7,844Nov 27, 2024Updated last year
- Tools for merging pretrained large language models.☆6,842Feb 28, 2026Updated last week
- The official implementation of Self-Play Fine-Tuning (SPIN)☆1,235May 8, 2024Updated last year
- Official repo for VGen: a holistic video generation ecosystem for video generation building on diffusion models☆3,154Jan 10, 2025Updated last year
- Fast and memory-efficient exact attention☆22,460Updated this week
- Official PyTorch Implementation of "Scalable Diffusion Models with Transformers"☆8,393May 31, 2024Updated last year
- A curated list of reinforcement learning with human feedback resources (continually updated)☆4,317Dec 9, 2025Updated 3 months ago
- Scaling Data-Constrained Language Models☆342Jun 28, 2025Updated 8 months ago
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities☆22,040Jan 23, 2026Updated last month
- Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.☆17,973Nov 3, 2025Updated 4 months ago
- Code and documents of LongLoRA and LongAlpaca (ICLR 2024 Oral)☆2,694Aug 14, 2024Updated last year
- PyTorch native post-training library☆5,697Updated this week
- Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization.☆10,358Jul 1, 2024Updated last year