evilsocket / cakeLinks
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
☆2,887Updated last year
Alternatives and similar repositories for cake
Users that are interested in cake are comparing it to the libraries listed below
Sorting:
- Blazingly fast LLM inference.☆6,171Updated last week
 - Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,713Updated last week
 - The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,534Updated 2 weeks ago
 - Run Mixtral-8x7B models in Colab or consumer desktops☆2,327Updated last year
 - SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.☆1,764Updated 4 months ago
 - Deep learning at the speed of light.☆2,586Updated this week
 - tiny vision language model☆8,863Updated last month
 - Local AI API Platform☆2,760Updated 3 months ago
 - Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚☆32,186Updated last week
 - Together Mixture-Of-Agents (MoA) – 65.1% on AlpacaEval with OSS models☆2,833Updated 9 months ago
 - AI app store powered by 24/7 desktop history. open source | 100% local | dev friendly | 24/7 screen, mic recording☆15,869Updated 2 months ago
 - High-speed Large Language Model Serving for Local Deployment☆8,374Updated 3 months ago
 - Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs☆3,526Updated 5 months ago
 - Run PyTorch LLMs locally on servers, desktop and mobile☆3,617Updated last month
 - A blazing fast inference solution for text embeddings models☆4,131Updated 3 weeks ago
 - A lightweight library for portable low-level GPU computation using WebGPU.☆3,914Updated 3 weeks ago
 - CoreNet: A library for training deep neural networks☆7,023Updated 3 weeks ago
 - An open source AI wearable device that captures what you say and hear in the real world and then transcribes and stores it on your own se…☆3,305Updated last year
 - Implementation for MatMul-free LM.☆3,034Updated 3 months ago
 - LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a ch…☆5,953Updated 6 months ago
 - AICI: Prompts as (Wasm) Programs☆2,052Updated 9 months ago
 - Making the community's best AI chat models available to everyone.☆1,983Updated 9 months ago
 - Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audi…☆9,043Updated this week
 - A fast llama2 decoder in pure Rust.☆1,051Updated last year
 - Distributed Training Over-The-Internet☆963Updated 2 weeks ago
 - Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/☆9,327Updated 5 months ago
 - g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains☆4,225Updated last month
 - llama and other large language models on iOS and MacOS offline using GGML library.☆1,904Updated last month
 - INT4/INT5/INT8 and FP16 inference on CPU for RWKV language model☆1,550Updated 7 months ago
 - Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization☆1,380Updated 11 months ago