An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs
☆686Mar 17, 2026Updated last week
Alternatives and similar repositories for exllamav3
Users that are interested in exllamav3 are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- The official API server for Exllama. OAI compatible, lightweight, and fast.☆1,158Updated this week
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,468Mar 4, 2026Updated 2 weeks ago
- ☆93Dec 9, 2025Updated 3 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆1,846Updated this week
- Web UI for ExLlamaV2☆510Feb 5, 2025Updated last year
- Produce your own Dynamic 3.0 Quants and achieve optimum accuracy & SOTA quantization performance! Input your VRAM and RAM and the toolcha…☆82Updated this week
- ☆166Jun 22, 2025Updated 9 months ago
- Large-scale LLM inference engine☆1,677Mar 12, 2026Updated last week
- ☆72Jun 20, 2025Updated 9 months ago
- Prompt Jinja2 templates for LLMs☆35Jul 9, 2025Updated 8 months ago
- Croco.Cpp is fork of KoboldCPP infering GGML/GGUF models on CPU/Cuda with KoboldAI's UI. It's powered partly by IK_LLama.cpp, and compati…☆161Mar 12, 2026Updated last week
- Yet Another (LLM) Web UI, made with Gemini☆12Dec 25, 2024Updated last year
- ☆64Jul 10, 2025Updated 8 months ago
- Transplants vocabulary between language models, enabling the creation of draft models for speculative decoding WITHOUT retraining.☆50Oct 29, 2025Updated 4 months ago
- A simple Gradio WebUI for loading/unloading models and loras in tabbyAPI.☆20Nov 21, 2024Updated last year
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆2,913Sep 30, 2023Updated 2 years ago
- ik_llama.cpp's Thireus fork with release builds for macOS/Windows/Ubuntu CPU, Vulkan and CUDA☆74Updated this week
- A multimodal, function calling powered LLM webui.☆215Sep 23, 2024Updated last year
- LLM model quantization (compression) toolkit with hw acceleration support for Nvidia CUDA, AMD ROCm, Intel XPU and Intel/AMD/Apple CPU vi…☆1,061Updated this week
- REAP: Router-weighted Expert Activation Pruning for SMoE compression☆292Updated this week
- Modified Beam Search with periodical restart☆12Sep 12, 2024Updated last year
- An extension to Oobabooga to add a simple memory function for chat☆25Jun 5, 2023Updated 2 years ago
- ☆54Oct 10, 2025Updated 5 months ago
- Run GGUF models easily with a KoboldAI UI. One File. Zero Install.☆9,793Updated this week
- An unsupervised model merging algorithm for Transformers-based language models.☆108Apr 29, 2024Updated last year
- Official implementation of Half-Quadratic Quantization (HQQ)☆919Feb 26, 2026Updated 3 weeks ago
- LLM Frontend in a single html file☆709Dec 27, 2025Updated 2 months ago
- 🎯An accuracy-first, highly efficient quantization toolkit for LLMs, designed to minimize quality degradation across Weight-Only Quantiza…☆914Updated this week
- Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc☆2,868Updated this week
- Transformers-compatible library for applying various compression algorithms to LLMs for optimized deployment with vLLM☆2,891Updated this week
- An OpenAI API compatible LLM inference server based on ExLlamaV2.☆25Feb 9, 2024Updated 2 years ago
- The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.☆14Mar 30, 2024Updated last year
- Optimizing inference proxy for LLMs☆3,389Updated this week
- Customizable implementation of the self-instruct paper.☆1,052Mar 7, 2024Updated 2 years ago
- A fast batching API to serve LLM models☆189Apr 26, 2024Updated last year
- ☆134Mar 14, 2026Updated last week
- A open webui function for better R1 experience☆78Mar 7, 2025Updated last year
- A stable, fast and easy-to-use inference library with a focus on a sync-to-async API☆48Sep 26, 2024Updated last year
- ☆110Aug 21, 2025Updated 7 months ago