evilsocket / cakeLinks
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
☆2,882Updated 9 months ago
Alternatives and similar repositories for cake
Users that are interested in cake are comparing it to the libraries listed below
Sorting:
- Blazingly fast LLM inference.☆5,996Updated this week
- Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,269Updated this week
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,465Updated this week
- Deep learning at the speed of light.☆2,079Updated last week
- Fast and accurate automatic speech recognition (ASR) for edge devices☆2,805Updated 3 months ago
- Together Mixture-Of-Agents (MoA) – 65.1% on AlpacaEval with OSS models☆2,796Updated 7 months ago
- LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a ch…☆5,948Updated 3 months ago
- Run PyTorch LLMs locally on servers, desktop and mobile☆3,603Updated last month
- Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audi…☆8,760Updated this week
- official repository of aiXcoder-7B Code Large Language Model☆2,271Updated last month
- An open source AI wearable device that captures what you say and hear in the real world and then transcribes and stores it on your own se…☆3,240Updated last year
- SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.☆1,752Updated last month
- g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains☆4,227Updated 6 months ago
- Local AI API Platform☆2,766Updated last month
- A vector search SQLite extension that runs anywhere!☆5,969Updated 6 months ago
- 🔍 AI search engine - self-host with local or cloud LLMs☆3,397Updated 10 months ago
- Local realtime voice AI☆2,347Updated 5 months ago
- AICI: Prompts as (Wasm) Programs☆2,044Updated 6 months ago
- A lightweight library for portable low-level GPU computation using WebGPU.☆3,899Updated 3 weeks ago
- ⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Pl…☆2,170Updated 10 months ago
- A cross-platform browser ML framework.☆710Updated 8 months ago
- AI app store powered by 24/7 desktop history. open source | 100% local | dev friendly | 24/7 screen, mic recording☆15,411Updated this week
- A fast llama2 decoder in pure Rust.☆1,054Updated last year
- prima.cpp: Speeding up 70B-scale LLM inference on low-resource everyday home clusters☆997Updated 3 weeks ago
- Llama-3 agents that can browse the web by following instructions and talking to you☆1,409Updated 8 months ago
- A self-organizing file system with llama 3☆5,356Updated this week
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆788Updated 2 weeks ago
- A nanoGPT pipeline packed in a spreadsheet☆2,120Updated last year
- High-speed Large Language Model Serving for Local Deployment☆8,301Updated last week
- A blazing fast inference solution for text embeddings models☆3,871Updated this week