huggingface / peft
π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
β15,839Updated this week
Related projects: β
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β10,327Updated last month
- Fast and memory-efficient exact attentionβ13,401Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMsβ9,906Updated 3 months ago
- Train transformer language models with reinforcement learning.β9,288Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMsβ26,822Updated this week
- Latest Advances on Multimodal Large Language Modelsβ11,722Updated this week
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.β19,294Updated last month
- Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom dataseβ¦β11,582Updated last week
- RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the bestβ¦β12,397Updated 2 weeks ago
- LAVIS - A One-stop Library for Language-Vision Intelligenceβ9,663Updated 3 weeks ago
- Accessible large language models via k-bit quantization for PyTorch.β6,029Updated this week
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β7,687Updated this week
- Large Language Model Text Generation Inferenceβ8,762Updated this week
- Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)β30,812Updated this week
- Instruct-tune LLaMA on consumer hardwareβ18,537Updated last month
- Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adβ¦β5,958Updated last week
- A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)β9,282Updated 3 months ago
- ImageBind One Embedding Space to Bind Them Allβ8,221Updated last month
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesβ19,545Updated 3 weeks ago
- [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parametersβ5,691Updated 6 months ago
- Ongoing research training transformer models at scaleβ9,949Updated this week
- An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.β8,212Updated last week
- Retrieval and Retrieval-augmented LLMsβ6,824Updated this week
- A framework for few-shot evaluation of language models.β6,426Updated this week
- Code and documentation to train Stanford's Alpaca models, and generate the data.β29,346Updated 2 months ago
- The official GitHub page for the survey paper "A Survey of Large Language Models".β10,016Updated 3 weeks ago
- An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.β36,446Updated this week
- An open source implementation of CLIP.β9,782Updated last month
- Example models using DeepSpeedβ5,987Updated this week
- Hackable and optimized Transformers building blocks, supporting a composable construction.β8,351Updated this week