QLoRA with Enhanced Multi GPU Support
☆38Aug 8, 2023Updated 2 years ago
Alternatives and similar repositories for qlora-multi-gpu
Users that are interested in qlora-multi-gpu are comparing it to the libraries listed below
Sorting:
- ☆13Aug 23, 2024Updated last year
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆101May 30, 2023Updated 2 years ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆83Sep 10, 2023Updated 2 years ago
- Official PyTorch implementation of QA-LoRA☆145Mar 13, 2024Updated last year
- Cuda extensions for PyTorch☆12Dec 2, 2025Updated 3 months ago
- A chat implementation for FastHTML☆11Sep 14, 2025Updated 5 months ago
- ☆13Feb 18, 2024Updated 2 years ago
- ☆18Apr 3, 2023Updated 2 years ago
- Build modern UIs in Jupyter with Python☆12Dec 28, 2022Updated 3 years ago
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Aug 25, 2023Updated 2 years ago
- A public implementation of the ReLoRA pretraining method, built on Lightning-AI's Pytorch Lightning suite.☆34Mar 2, 2024Updated 2 years ago
- Datasets and code from our paper, where we use machine learning to predict if ChatGPT will refuse a given prompt.☆38Sep 23, 2023Updated 2 years ago
- ☆63Sep 23, 2024Updated last year
- Low-Rank adapter extraction for fine-tuned transformers models☆180May 2, 2024Updated last year
- A collection of utilities for FastHTML projects.☆14Oct 23, 2024Updated last year
- Full finetuning of large language models without large memory requirements☆94Sep 22, 2025Updated 5 months ago
- Generate textbook-quality synthetic LLM pretraining data☆509Oct 19, 2023Updated 2 years ago
- Official Documentation for DSPy Library☆21Feb 26, 2026Updated last week
- ☆17Feb 16, 2024Updated 2 years ago
- A library for squeakily cleaning and filtering language datasets.☆50Jul 10, 2023Updated 2 years ago
- Simplex Random Feature attention, in PyTorch☆76Oct 10, 2023Updated 2 years ago
- Positional Skip-wise Training for Efficient Context Window Extension of LLMs to Extremely Length (ICLR 2024)☆209May 20, 2024Updated last year
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆233Oct 31, 2024Updated last year
- Code repo for "Model-Generated Pretraining Signals Improves Zero-Shot Generalization of Text-to-Text Transformers" (ACL 2023)☆22Nov 1, 2023Updated 2 years ago
- ☆74Sep 5, 2023Updated 2 years ago
- QLoRA: Efficient Finetuning of Quantized LLMs☆79Apr 10, 2024Updated last year
- Library for extremely fast HTML generation from Python☆27Oct 24, 2024Updated last year
- Visual Studio Code extension to convert HTML to FastHTML FT☆22Feb 21, 2025Updated last year
- A stable, fast and easy-to-use inference library with a focus on a sync-to-async API☆48Sep 26, 2024Updated last year
- Code for PHATGOOSE introduced in "Learning to Route Among Specialized Experts for Zero-Shot Generalization"☆91Feb 27, 2024Updated 2 years ago
- Utilities for Training Very Large Models☆58Sep 25, 2024Updated last year
- QLoRA for Masked Language Modeling☆23Sep 11, 2023Updated 2 years ago
- Multipack distributed sampler for fast padding-free training of LLMs☆204Aug 10, 2024Updated last year
- ☆21Mar 3, 2025Updated last year
- ☆25Feb 18, 2024Updated 2 years ago
- Helpers and such for working with Lambda Cloud☆52Nov 7, 2023Updated 2 years ago
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytes…☆145Oct 17, 2023Updated 2 years ago
- Yaraa (Yet Another Rag Automation Attempt) is a library that tackles the boring aspects of managing Rag pipelines, so you don't have to.☆26Sep 5, 2024Updated last year
- ModuleFormer is a MoE-based architecture that includes two different types of experts: stick-breaking attention heads and feedforward exp…☆226Sep 18, 2025Updated 5 months ago