microsoft / LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
β10,327Updated last month
Related projects: β
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β15,839Updated this week
- Fast and memory-efficient exact attentionβ13,401Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMsβ9,906Updated 3 months ago
- Train transformer language models with reinforcement learning.β9,288Updated this week
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.β19,294Updated last month
- LAVIS - A One-stop Library for Language-Vision Intelligenceβ9,663Updated 3 weeks ago
- RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the bestβ¦β12,397Updated 2 weeks ago
- ImageBind One Embedding Space to Bind Them Allβ8,221Updated last month
- Hackable and optimized Transformers building blocks, supporting a composable construction.β8,351Updated this week
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β7,687Updated this week
- Accessible large language models via k-bit quantization for PyTorch.β6,029Updated this week
- Instruct-tune LLaMA on consumer hardwareβ18,537Updated last month
- Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adβ¦β5,958Updated last week
- Scripts for fine-tuning Meta Llama3 with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom dataseβ¦β11,582Updated last week
- An open source implementation of CLIP.β9,782Updated last month
- [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parametersβ5,691Updated 6 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMsβ26,822Updated this week
- Latest Advances on Multimodal Large Language Modelsβ11,722Updated this week
- π€ Diffusers: State-of-the-art diffusion models for image and audio generation in PyTorch and FLAX.β25,168Updated this week
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesβ19,545Updated 3 weeks ago
- Large Language Model Text Generation Inferenceβ8,762Updated this week
- A collection of libraries to optimise AI model performancesβ8,373Updated last month
- Code and documentation to train Stanford's Alpaca models, and generate the data.β29,346Updated 2 months ago
- Example models using DeepSpeedβ5,987Updated this week
- tiktoken is a fast BPE tokeniser for use with OpenAI's models.β11,796Updated last month
- Ongoing research training transformer models at scaleβ9,949Updated this week
- Official repo for consistency models.β6,073Updated 5 months ago
- LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMathβ9,194Updated last month
- PyTorch code and models for the DINOv2 self-supervised learning method.β8,791Updated last month
- A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)β9,282Updated 3 months ago