rmihaylov / mpttuneView external linksLinks
Tune MPTs
☆84Jun 17, 2023Updated 2 years ago
Alternatives and similar repositories for mpttune
Users that are interested in mpttune are comparing it to the libraries listed below
Sorting:
- A repo for finetuning MPT using LoRA. It is currently configured to work with the Alpaca dataset from Stanford but can easily be adapted …☆18Jun 12, 2023Updated 2 years ago
- Tune any FALCON in 4-bit☆463Sep 1, 2023Updated 2 years ago
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆104May 20, 2025Updated 8 months ago
- ☆17Jun 19, 2023Updated 2 years ago
- Extism Plug-in Development Kit (PDK) for C☆14Oct 22, 2024Updated last year
- ☆553Updated this week
- Trying to deconstruct RWKV in understandable terms☆14May 6, 2023Updated 2 years ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31May 22, 2024Updated last year
- Patch for MPT-7B which allows using and training a LoRA☆58May 20, 2023Updated 2 years ago
- ☆13Jan 27, 2019Updated 7 years ago
- Optimizing bit-level Jaccard Index and Population Counts for large-scale quantized Vector Search via Harley-Seal CSA and Lookup Tables☆21May 18, 2025Updated 8 months ago
- Run fast LLM Inference using Llama.cpp in Python☆19Jan 3, 2024Updated 2 years ago
- Generate usage documentation in Markdown format from Python scripts using argparse☆17Jan 14, 2026Updated 3 weeks ago
- ☆21May 27, 2023Updated 2 years ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Jun 21, 2023Updated 2 years ago
- tinygrad port of the RWKV large language model.☆45Mar 9, 2025Updated 11 months ago
- Finetuning Large Language Models on One Consumer GPU in 2 Bits☆734May 25, 2024Updated last year
- 5X faster 60% less memory QLoRA finetuning☆21May 28, 2024Updated last year
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆59Oct 18, 2025Updated 3 months ago
- ☆56Jun 26, 2025Updated 7 months ago
- Implementation of the Mamba SSM with hf_integration.☆55Aug 31, 2024Updated last year
- Convert all of libgen to high quality markdown☆254Dec 13, 2023Updated 2 years ago
- Prune transformer layers☆74May 30, 2024Updated last year
- A large-scale information-rich web dataset, featuring millions of real clicked query-document labels☆346Dec 16, 2024Updated last year
- ☆27Aug 30, 2023Updated 2 years ago
- ☆27Jan 8, 2024Updated 2 years ago
- Code for the paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot" with LLaMA implementation.☆71Mar 30, 2023Updated 2 years ago
- Example ML projects that use the Determined library.☆32Sep 11, 2024Updated last year
- Experimental sampler to make LLMs more creative☆31Aug 2, 2023Updated 2 years ago
- Code for fine-tuning Platypus fam LLMs using LoRA☆630Feb 4, 2024Updated 2 years ago
- ☆39Aug 1, 2025Updated 6 months ago
- [MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration☆3,436Jul 17, 2025Updated 6 months ago
- Load local LLMs effortlessly in a Jupyter notebook for testing purposes alongside Langchain or other agents. Contains Oobagooga and Kobol…☆213Jun 25, 2023Updated 2 years ago
- OpenAlpaca: A Fully Open-Source Instruction-Following Model Based On OpenLLaMA☆304Jun 13, 2023Updated 2 years ago
- Official repository for LongChat and LongEval☆534May 24, 2024Updated last year
- ☆78Dec 26, 2023Updated 2 years ago
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆83Sep 10, 2023Updated 2 years ago
- ☆535Dec 1, 2023Updated 2 years ago
- Customizable implementation of the self-instruct paper.☆1,050Mar 7, 2024Updated last year