evilsocket / cakeLinks
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
☆2,888Updated last year
Alternatives and similar repositories for cake
Users that are interested in cake are comparing it to the libraries listed below
Sorting:
- Blazingly fast LLM inference.☆6,230Updated last week
- Run PyTorch LLMs locally on servers, desktop and mobile☆3,617Updated 2 months ago
- High-speed Large Language Model Serving for Local Deployment☆8,388Updated 3 months ago
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,538Updated last month
- Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,743Updated 3 weeks ago
- CoreNet: A library for training deep neural networks☆7,024Updated last month
- LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a ch…☆5,958Updated 6 months ago
- SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.☆1,779Updated 5 months ago
- A minimal GPU design in Verilog to learn how GPUs work from the ground up☆8,894Updated last year
- On-device AI across mobile, embedded and edge for PyTorch☆3,531Updated this week
- Fast and accurate automatic speech recognition (ASR) for edge devices☆2,968Updated this week
- Implementation for MatMul-free LM.☆3,037Updated 4 months ago
- lightweight, standalone C++ inference engine for Google's Gemma models.☆6,615Updated this week
- AI app store powered by 24/7 desktop history. open source | 100% local | dev friendly | 24/7 screen, mic recording☆15,979Updated 2 months ago
- Enchanted is iOS and macOS app for chatting with private self hosted language models such as Llama2, Mistral or Vicuna using Ollama.☆5,722Updated 8 months ago
- AIOS: AI Agent Operating System☆4,801Updated this week
- Distributed Training Over-The-Internet☆963Updated last month
- tiny vision language model☆8,917Updated last week
- Deep learning at the speed of light.☆2,632Updated last week
- Together Mixture-Of-Agents (MoA) – 65.1% on AlpacaEval with OSS models☆2,837Updated 10 months ago
- Local AI API Platform☆2,764Updated 4 months ago
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆939Updated last month
- A passive recording project allows you to have complete control over your data. Automatically take screenshots of all your screens, index…☆1,311Updated 2 months ago
- g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains☆4,223Updated 2 months ago
- llama and other large language models on iOS and MacOS offline using GGML library.☆1,915Updated 2 months ago
- Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization☆1,380Updated 11 months ago
- Local realtime voice AI☆2,378Updated 8 months ago
- A fast llama2 decoder in pure Rust.☆1,056Updated last year
- A vector search SQLite extension that runs anywhere!☆6,412Updated 9 months ago
- AICI: Prompts as (Wasm) Programs☆2,054Updated 10 months ago