gpustack / llama-box
LM inference server implementation based on *.cpp.
☆185Updated this week
Alternatives and similar repositories for llama-box
Users that are interested in llama-box are comparing it to the libraries listed below
Sorting:
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆161Updated this week
- A text-to-speech and speech-to-text server compatible with the OpenAI API, supporting Whisper, FunASR, Bark, and CosyVoice backends.☆109Updated 3 weeks ago
- xllamacpp - a Python wrapper of llama.cpp☆36Updated last week
- ☆88Updated 2 months ago
- run DeepSeek-R1 GGUFs on KTransformers☆226Updated 2 months ago
- Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)☆583Updated last week
- automatically quant GGUF models☆174Updated last week
- ☆142Updated 2 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆439Updated this week
- Production ready LLM model compression/quantization toolkit with hw accelerated inference support for both cpu/gpu via HF, vLLM, and SGLa…☆537Updated this week
- Open Source Text Embedding Models with OpenAI Compatible API☆153Updated 10 months ago
- Port of Facebook's LLaMA model in C/C++☆52Updated 2 weeks ago
- gpt_server是一个用于生产级部署LLMs、Embedding、Reranker、ASR和TTS的开源框架。☆177Updated this week
- CPU inference for the DeepSeek family of large language models in C++☆294Updated this week
- DashInfer is a native LLM inference engine aiming to deliver industry-leading performance atop various hardware architectures, including …☆245Updated this week
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆253Updated 2 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆131Updated 10 months ago
- ☆202Updated 3 weeks ago
- LongCite: Enabling LLMs to Generate Fine-grained Citations in Long-context QA☆489Updated 4 months ago
- Mixture-of-Experts (MoE) Language Model☆186Updated 8 months ago
- The latest graphrag interface is used, using the local ollama to provide the LLM interface.Support for using the pip installation☆147Updated 7 months ago
- Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).☆243Updated last year
- LLM inference in C/C++☆76Updated this week
- Lightweight Inference server for OpenVINO☆165Updated this week
- C++ implementation of Qwen-LM☆586Updated 5 months ago
- Port of Facebook's LLaMA model in C/C++☆93Updated this week
- A Pure Rust based LLM (Any LLM based MLLM such as Spark-TTS) Inference Engine, powering by Candle framework.☆106Updated last month
- 研究GOT-OCR-项目落地加速,不限语言☆60Updated 6 months ago
- ☆59Updated last year
- Auto Thinking Mode switch for Qwen3 in Open webui☆50Updated this week