evilsocket / cakeLinks
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
☆2,899Updated last year
Alternatives and similar repositories for cake
Users that are interested in cake are comparing it to the libraries listed below
Sorting:
- Blazingly fast LLM inference.☆6,379Updated this week
- Distributed LLM inference. Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,804Updated last week
- SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.☆1,800Updated 3 weeks ago
- Local AI API Platform☆2,761Updated 6 months ago
- Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audi…☆9,333Updated last week
- Instant, controllable, local pre-trained AI models in Rust☆2,122Updated last week
- Run PyTorch LLMs locally on servers, desktop and mobile☆3,625Updated 4 months ago
- A vector search SQLite extension that runs anywhere!☆6,723Updated last year
- A fast llama2 decoder in pure Rust.☆1,060Updated 2 years ago
- Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚☆40,273Updated last week
- AICI: Prompts as (Wasm) Programs☆2,063Updated last year
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆578Updated this week
- g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains☆4,219Updated 3 weeks ago
- High-speed Large Language Model Serving for Local Deployment☆8,591Updated 5 months ago
- Minimal LLM inference in Rust☆1,028Updated last year
- A blazing fast inference solution for text embeddings models☆4,417Updated this week
- Local realtime voice AI☆2,423Updated 2 months ago
- ☆3,497Updated last year
- A cross-platform browser ML framework.☆742Updated last year
- Calculate token/s & GPU memory requirement for any LLM. Supports llama.cpp/ggml/bnb/QLoRA quantization☆1,388Updated last year
- [Unmaintained, see README] An ecosystem of Rust libraries for working with large language models☆6,146Updated last year
- Distributed Training Over-The-Internet☆975Updated 3 months ago
- Ingest, parse, and optimize any data format ➡️ from documents to multimedia ➡️ for enhanced compatibility with GenAI frameworks☆6,781Updated last month
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,583Updated last month
- Deep learning at the speed of light.☆2,738Updated this week
- prime is a framework for efficient, globally distributed training of AI models over the internet.☆850Updated 2 months ago
- A high-performance inference engine for AI models☆1,418Updated this week
- Fast and accurate automatic speech recognition (ASR) for edge devices☆3,096Updated 2 months ago
- Large Action Model framework to develop AI Web Agents☆6,275Updated last year
- Run Mixtral-8x7B models in Colab or consumer desktops☆2,326Updated last year