iuliaturc / gguf-docsLinks
Docs for GGUF quantization (unofficial)
☆319Updated 4 months ago
Alternatives and similar repositories for gguf-docs
Users that are interested in gguf-docs are comparing it to the libraries listed below
Sorting:
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆578Updated 2 weeks ago
- InferX: Inference as a Service Platform☆139Updated this week
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆143Updated 4 months ago
- Guaranteed Structured Output from any Language Model via Hierarchical State Machines☆145Updated last month
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆247Updated 3 weeks ago
- ☆289Updated 3 weeks ago
- A little(lil) Language Model (LM). A tiny reproduction of LLaMA 3's model architecture.☆52Updated 7 months ago
- Enhancing LLMs with LoRA☆176Updated last month
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- llama.cpp fork with additional SOTA quants and improved performance☆1,341Updated this week
- Fast parallel LLM inference for MLX☆233Updated last year
- ☆209Updated 2 months ago
- A Conversational Speech Generation Model with Gradio UI and OpenAI compatible API. UI and API support CUDA, MLX and CPU devices.☆208Updated 6 months ago
- ☆399Updated 2 weeks ago
- Welcome to the official repository of SINQ! A novel, fast and high-quality quantization method designed to make any Large Language Model …☆578Updated this week
- Sparse Inferencing for transformer based LLMs☆213Updated 3 months ago
- ☆135Updated 6 months ago
- automatically quant GGUF models☆214Updated last month
- Transplants vocabulary between language models, enabling the creation of draft models for speculative decoding WITHOUT retraining.☆46Updated last month
- API Server for Transformer Lab☆80Updated last week
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆278Updated 3 months ago
- ☆28Updated 5 months ago
- Live-bending a foundation model’s output at neural network level.☆270Updated 7 months ago
- ☆703Updated last week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆84Updated last month
- A platform to self-host AI on easy mode☆177Updated last week
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆132Updated 3 weeks ago
- LLM Inference on consumer devices☆125Updated 8 months ago
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆85Updated last year
- AI management tool☆121Updated last year