microsoft / LoRA
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
โ11,847Updated 4 months ago
Alternatives and similar repositories for LoRA:
Users that are interested in LoRA are comparing it to the libraries listed below
- ๐ค PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.โ18,274Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMsโ10,411Updated 10 months ago
- Train transformer language models with reinforcement learning.โ13,559Updated this week
- Fast and memory-efficient exact attentionโ17,192Updated this week
- ๐ A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iโฆโ8,673Updated this week
- Accessible large language models via k-bit quantization for PyTorch.โ6,972Updated this week
- [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parametersโ5,868Updated last year
- LAVIS - A One-stop Library for Language-Vision Intelligenceโ10,516Updated 5 months ago
- [NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.โ22,354Updated 8 months ago
- Instruct-tune LLaMA on consumer hardwareโ18,902Updated 9 months ago
- Code and documentation to train Stanford's Alpaca models, and generate the data.โ29,972Updated 9 months ago
- ๐ค Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch and FLAX.โ28,835Updated this week
- a state-of-the-art-level open visual language model | ๅคๆจกๆ้ข่ฎญ็ปๆจกๅโ6,507Updated 11 months ago
- An open source implementation of CLIP.โ11,654Updated last week
- An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.โ38,502Updated 3 weeks ago
- A high-throughput and memory-efficient inference and serving engine for LLMsโ46,456Updated this week
- [ICLR 2024] Efficient Streaming Language Models with Attention Sinksโ6,872Updated 9 months ago
- Hackable and optimized Transformers building blocks, supporting a composable construction.โ9,435Updated last week
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesโ21,171Updated 2 months ago
- Aligning pretrained language models with instruction data generated by themselves.โ4,359Updated 2 years ago
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)โฆโ13,564Updated this week
- Ongoing research training transformer models at scaleโ12,261Updated this week
- An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.โ8,411Updated this week
- ImageBind One Embedding Space to Bind Them Allโ8,627Updated 9 months ago
- Retrieval and Retrieval-augmented LLMsโ9,524Updated 3 weeks ago
- LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMathโ9,391Updated 9 months ago
- GLM-130B: An Open Bilingual Pre-Trained Model (ICLR 2023)โ7,680Updated last year
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.โ4,826Updated 3 weeks ago
- The official repo of Qwen-VL (้ไนๅ้ฎ-VL) chat & pretrained large vision language model proposed by Alibaba Cloud.โ5,842Updated 8 months ago
- Instruction Tuning with GPT-4โ4,301Updated last year