fredlas / optimize_llamacpp_nglLinks
empirically chooses -ngl param for llama.cpp
☆17Updated 10 months ago
Alternatives and similar repositories for optimize_llamacpp_ngl
Users that are interested in optimize_llamacpp_ngl are comparing it to the libraries listed below
Sorting:
- Controllable Language Model Interactions in TypeScript☆10Updated last year
- Simple LLM inference server☆20Updated last year
- ☆23Updated last year
- BlinkDL's RWKV-v4 running in the browser☆48Updated 2 years ago
- Python package wrapping llama.cpp for on-device LLM inference☆100Updated 3 months ago
- Port of Facebook's LLaMA model in C/C++☆21Updated 2 years ago
- Generate Structured JSON with probs from Language Models☆17Updated 10 months ago
- The hearth of The Pulsar App, fast, secure and shared inference with modern UI☆59Updated last year
- Accepts a Hugging Face model URL, automatically downloads and quantizes it using Bits and Bytes.☆38Updated last year
- Yet Another (LLM) Web UI, made with Gemini☆12Updated last year
- Modified Beam Search with periodical restart☆12Updated last year
- An OpenAI API compatible LLM inference server based on ExLlamaV2.☆25Updated 2 years ago
- Game Companion AI is an advanced application designed to enhance the gaming experience by providing real-time analysis and interpretation…