huggingface / trlLinks
Train transformer language models with reinforcement learning.
β17,115Updated this week
Alternatives and similar repositories for trl
Users that are interested in trl are comparing it to the libraries listed below
Sorting:
- An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & TIS & vLLM & Ray & Async RL)β8,851Updated this week
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β20,502Updated this week
- A framework for few-shot evaluation of language models.β11,246Updated this week
- verl: Volcano Engine Reinforcement Learning for LLMsβ18,535Updated this week
- Fast and memory-efficient exact attentionβ21,773Updated this week
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)β4,738Updated 2 years ago
- Tools for merging pretrained large language models.β6,696Updated 3 weeks ago
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β9,461Updated last week
- Accessible large language models via k-bit quantization for PyTorch.β7,912Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMsβ10,815Updated last year
- Ongoing research training transformer models at scaleβ14,939Updated last week
- General technology for enabling AI capabilities w/ LLMs and MLLMsβ4,262Updated last month
- Retrieval and Retrieval-augmented LLMsβ11,187Updated last month
- Large Language Model Text Generation Inferenceβ10,731Updated 2 weeks ago
- Reference implementation for DPO (Direct Preference Optimization)β2,832Updated last year
- Aligning pretrained language models with instruction data generated by themselves.β4,562Updated 2 years ago
- Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adβ¦β6,093Updated 6 months ago
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β13,169Updated last year
- Robust recipes to align language models with human and AI preferencesβ5,481Updated 4 months ago
- PyTorch native post-training libraryβ5,649Updated this week
- Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We alsβ¦β18,148Updated 2 months ago
- Go ahead and axolotl questionsβ11,138Updated this week
- [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parametersβ5,936Updated last year
- A curated list of reinforcement learning with human feedback resources (continually updated)β4,270Updated last month
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.β5,023Updated 9 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMsβ68,244Updated this week
- AllenAI's post-training codebaseβ3,538Updated this week
- [ICLR 2024] Efficient Streaming Language Models with Attention Sinksβ7,172Updated last year
- LMDeploy is a toolkit for compressing, deploying, and serving LLMs.β7,544Updated this week
- The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.β8,869Updated last year