Tune MPTs
☆84Jun 17, 2023Updated 2 years ago
Alternatives and similar repositories for mpttune
Users that are interested in mpttune are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- A repo for finetuning MPT using LoRA. It is currently configured to work with the Alpaca dataset from Stanford but can easily be adapted …☆18Jun 12, 2023Updated 2 years ago
- Command-line script for inferencing from models such as MPT-7B-Chat☆100Jul 6, 2023Updated 2 years ago
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆105Updated this week
- Tune any FALCON in 4-bit☆462Sep 1, 2023Updated 2 years ago
- Patch for MPT-7B which allows using and training a LoRA☆58May 20, 2023Updated 2 years ago
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- ☆553Feb 8, 2026Updated 2 months ago
- ☆12Jun 18, 2019Updated 6 years ago
- ☆17Jun 19, 2023Updated 2 years ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆31May 22, 2024Updated last year
- Trying to deconstruct RWKV in understandable terms☆14May 6, 2023Updated 2 years ago
- A Next.js chat app to use Llama 2 locally using node-llama-cpp☆12Oct 27, 2024Updated last year
- Run fast LLM Inference using Llama.cpp in Python☆19Jan 3, 2024Updated 2 years ago
- Optimizing bit-level Jaccard Index and Population Counts for large-scale quantized Vector Search via Harley-Seal CSA and Lookup Tables☆22May 18, 2025Updated 10 months ago
- ☆16Jul 20, 2023Updated 2 years ago
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- tinygrad port of the RWKV large language model.☆45Mar 9, 2025Updated last year
- ☆15Sep 8, 2023Updated 2 years ago
- Code for the paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot" with LLaMA implementation.☆71Mar 30, 2023Updated 3 years ago
- Large Language Model Hosting Container☆92Updated this week
- Finetuning Large Language Models on One Consumer GPU in 2 Bits☆733May 25, 2024Updated last year
- data prep utilities for LLMs, using LLMs☆16Nov 7, 2023Updated 2 years ago
- [ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization☆717Aug 13, 2024Updated last year
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆60Oct 18, 2025Updated 5 months ago
- An ONNX converter script focused on embedding models☆33Jan 14, 2025Updated last year
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- 5X faster 60% less memory QLoRA finetuning☆21May 28, 2024Updated last year
- ☆37May 31, 2023Updated 2 years ago
- ☆12Oct 3, 2024Updated last year
- Paper Implementation of Self-Rewarding Language Models☆13Feb 1, 2024Updated 2 years ago
- ☆10Jul 25, 2023Updated 2 years ago
- YOLOX implemented by pytorch lightning, a simpler expression of pytorch☆11May 26, 2022Updated 3 years ago
- Port of Facebook's LLaMA model in C/C++☆21Nov 6, 2023Updated 2 years ago
- A repository for a Java implementation of AutoGPT. It is heavily inspired by AutoGPT if not a clone of some of its functionality, though …☆27May 9, 2023Updated 2 years ago
- Track the progress of LLM context utilisation☆55Apr 14, 2025Updated last year
- Managed Kubernetes at scale on DigitalOcean • AdDigitalOcean Kubernetes includes the control plane, bandwidth allowance, container registry, automatic updates, and more for free.
- A library for simplifying training with multi gpu setups in the HuggingFace / PyTorch ecosystem.☆16Mar 16, 2026Updated last month
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆38Jun 6, 2023Updated 2 years ago
- ☆56Jun 26, 2025Updated 9 months ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Jun 21, 2023Updated 2 years ago
- ☆27Jan 8, 2024Updated 2 years ago
- ☆78Dec 26, 2023Updated 2 years ago
- OpenAlpaca: A Fully Open-Source Instruction-Following Model Based On OpenLLaMA☆304Jun 13, 2023Updated 2 years ago