antirez / gguf-toolsLinks
GGUF implementation in C as a library and a tools CLI program
☆277Updated 6 months ago
Alternatives and similar repositories for gguf-tools
Users that are interested in gguf-tools are comparing it to the libraries listed below
Sorting:
- A minimalistic C++ Jinja templating engine for LLM chat templates☆163Updated 3 weeks ago
- Inference of Mamba models in pure C☆189Updated last year
- LLM-based code completion engine☆193Updated 6 months ago
- throwaway GPT inference☆140Updated last year
- An implementation of bucketMul LLM inference☆221Updated last year
- Python bindings for ggml☆142Updated 11 months ago
- LLaVA server (llama.cpp).☆181Updated last year
- GGML implementation of BERT model with Python bindings and quantization.☆56Updated last year
- Run GGML models with Kubernetes.☆173Updated last year
- A faithful clone of Karpathy's llama2.c (one file inference, zero dependency) but fully functional with LLaMA 3 8B base and instruct mode…☆129Updated last year
- Stateful load balancer custom-tailored for llama.cpp 🏓🦙☆800Updated this week
- ☆60Updated 11 months ago
- WebGPU LLM inference tuned by hand☆151Updated 2 years ago
- 1.58 Bit LLM on Apple Silicon using MLX☆217Updated last year
- C API for MLX☆121Updated 3 weeks ago
- Extend the original llama.cpp repo to support redpajama model.☆118Updated 11 months ago
- ggml implementation of BERT☆495Updated last year
- a small code base for training large models☆307Updated 3 months ago
- Port of MiniGPT4 in C++ (4bit, 5bit, 6bit, 8bit, 16bit CPU inference with GGML)☆568Updated last year
- ☆388Updated last week
- Fast parallel LLM inference for MLX☆204Updated last year
- Mistral7B playing DOOM☆133Updated last year
- Heirarchical Navigable Small Worlds☆98Updated 3 months ago
- CLIP inference in plain C/C++ with no extra dependencies☆514Updated last month
- run embeddings in MLX☆90Updated 10 months ago
- Pytorch script hot swap: Change code without unloading your LLM from VRAM☆126Updated 3 months ago
- SoTA Transformers with C-backend for fast inference on your CPU.☆309Updated last year
- ☆249Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attention☆119Updated last year
- Falcon LLM ggml framework with CPU and GPU support☆246Updated last year