Distributed inference for mobile, desktop and server.
☆3,027Apr 24, 2026Updated last week
Alternatives and similar repositories for cake
Users that are interested in cake are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Run frontier AI locally.☆44,096Updated this week
- Fast, flexible LLM inference☆7,074Apr 15, 2026Updated 2 weeks ago
- Minimalist ML framework for Rust☆20,082Apr 23, 2026Updated last week
- Claude Engineer is an interactive command-line interface (CLI) that leverages the power of Anthropic's Claude-3.5-Sonnet model to assist …☆11,167Dec 12, 2024Updated last year
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆644Apr 22, 2026Updated last week
- GPUs on demand by Runpod - Special Offer Available • AdRun AI, ML, and HPC workloads on powerful cloud GPUs—without limits or wasted spend. Deploy GPUs in under a minute and pay by the second.
- Burn is a next generation tensor library and Deep Learning Framework that doesn't compromise on flexibility, efficiency and portability.☆14,938Updated this week
- Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,908Apr 14, 2026Updated 2 weeks ago
- Run agents that work for you based on what you do. AI finally knows what you are doing☆18,405Updated this week
- [Unmaintained, see README] An ecosystem of Rust libraries for working with large language models☆6,152Jun 24, 2024Updated last year
- Universal memory layer for AI Agents☆54,199Updated this week
- Universal LLM Deployment Engine with ML Compilation☆22,517Apr 22, 2026Updated last week
- Distribute and run LLMs with a single file.☆24,274Apr 23, 2026Updated last week
- 🔍 An LLM-based Multi-agent Framework of Web Search Engine (like Perplexity.ai Pro and SearchGPT)☆6,851Jul 4, 2025Updated 9 months ago
- Qdrant - High-performance, massive-scale Vector Database and Vector Search Engine for the next generation of AI. Also available in the cl…☆30,799Updated this week
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- High-speed Large Language Model Serving for Local Deployment☆9,390Jan 24, 2026Updated 3 months ago
- LLM inference in C/C++☆106,639Updated this week
- A lightweight library for portable low-level GPU computation using WebGPU.☆3,967Oct 8, 2025Updated 6 months ago
- A library for building fast, reliable and evolvable network services.☆26,497Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆78,385Updated this week
- Web UI for training and running open models like Gemma 4, Qwen3.6, DeepSeek, gpt-oss locally.☆63,070Updated this week
- Ingest, parse, and optimize any data format ➡️ from documents to multimedia ➡️ for enhanced compatibility with GenAI frameworks☆6,812Dec 12, 2025Updated 4 months ago
- Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audi…☆10,070Apr 24, 2026Updated last week
- Apache OpenDAL: One Layer, All Storage.☆5,032Updated this week
- Deploy open-source AI quickly and easily - Special Bonus Offer • AdRunpod Hub is built for open source. One-click deployment and autoscaling endpoints without provisioning your own infrastructure.
- A Gemini 2.5 Flash Level MLLM for Vision, Speech, and Full-Duplex Multimodal Live Streaming on Your Phone☆24,460Updated this week
- A modular graph-based Retrieval-Augmented Generation (RAG) system☆32,491Updated this week
- Tantivy is a full-text search engine library inspired by Apache Lucene and written in Rust☆15,070Updated this week
- SGLang is a high-performance serving framework for large language models and multimodal models.☆26,397Updated this week
- An open-source RAG-based tool for chatting with your documents.☆25,310Apr 3, 2026Updated 3 weeks ago
- Self-hosted AI coding assistant☆33,473Mar 2, 2026Updated last month
- LLM training in simple, raw C/CUDA☆29,687Jun 26, 2025Updated 10 months ago
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,626Feb 8, 2026Updated 2 months ago
- Get up and running with Kimi-K2.5, GLM-5, MiniMax, DeepSeek, gpt-oss, Qwen, Gemma and other models.☆170,289Updated this week
- GPU virtual machines on DigitalOcean Gradient AI • AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- Inference Llama 2 in one file of pure C☆19,440Aug 6, 2024Updated last year
- 🌸 Run LLMs at home, BitTorrent-style. Fine-tuning and inference up to 10x faster than offloading☆10,086Sep 7, 2024Updated last year
- Open-source LLM/VLM load balancer and serving platform for self-hosting LLMs (and VLMs) at scale 🏓🦙 Alternative to projects like llm-d,…☆1,540Updated this week
- Together Mixture-Of-Agents (MoA) – 65.1% on AlpacaEval with OSS models☆2,890Jan 7, 2025Updated last year
- ⚙️🦀 Build modular and scalable LLM Applications in Rust☆7,053Apr 24, 2026Updated last week
- SOTA Open Source TTS☆29,922Apr 6, 2026Updated 3 weeks ago
- An LLM-powered knowledge curation system that researches a topic and generates a full-length report with citations.☆28,127Sep 30, 2025Updated 7 months ago