evilsocket / cakeLinks
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
☆2,859Updated 8 months ago
Alternatives and similar repositories for cake
Users that are interested in cake are comparing it to the libraries listed below
Sorting:
- Blazingly fast LLM inference.☆5,764Updated this week
- Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆2,189Updated last week
- SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.☆1,730Updated last week
- High-speed Large Language Model Serving for Local Deployment☆8,224Updated 4 months ago
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,431Updated last week
- 🔍 An LLM-based Multi-agent Framework of Web Search Engine (like Perplexity.ai Pro and SearchGPT)☆6,403Updated 5 months ago
- LLocalSearch is a completely locally running search aggregator using LLM Agents. The user can ask a question and the system will use a ch…☆5,926Updated last month
- Reverse Engineering: Decompiling Binary Code with Large Language Models☆5,710Updated this week
- CoreNet: A library for training deep neural networks☆7,016Updated last month
- A blazing fast inference solution for text embeddings models☆3,707Updated last week
- Stateful load balancer custom-tailored for llama.cpp 🏓🦙☆782Updated this week
- AIOS: AI Agent Operating System☆4,260Updated 2 weeks ago
- Moshi is a speech-text foundation model and full-duplex spoken dialogue framework. It uses Mimi, a state-of-the-art streaming neural audi…☆8,460Updated this week
- open-source multimodal large language model that can hear, talk while thinking. Featuring real-time end-to-end speech input and streaming…☆3,343Updated 7 months ago
- Speech To Speech: an effort for an open-sourced and modular GPT4-o☆4,078Updated 2 months ago
- Ingest, parse, and optimize any data format ➡️ from documents to multimedia ➡️ for enhanced compatibility with GenAI frameworks☆6,586Updated 2 weeks ago
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.☆1,347Updated 2 weeks ago
- Local realtime voice AI☆2,328Updated 3 months ago
- Virtual Machine for the Web☆13,172Updated last week
- g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains☆4,223Updated 4 months ago
- Run your own AI cluster at home with everyday devices 📱💻 🖥️⌚☆28,527Updated 3 months ago
- Fast and accurate automatic speech recognition (ASR) for edge devices☆2,757Updated last month
- LLaMA-Omni is a low-latency and high-quality end-to-end speech interaction model built upon Llama-3.1-8B-Instruct, aiming to achieve spee…☆2,942Updated last month
- MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.☆1,303Updated 2 months ago
- lightweight, standalone C++ inference engine for Google's Gemma models.☆6,471Updated this week
- A vector search SQLite extension that runs anywhere!☆5,791Updated 5 months ago
- Examples in the MLX framework☆7,555Updated 2 weeks ago
- AI app store powered by 24/7 desktop history. open source | 100% local | dev friendly | 24/7 screen, mic recording☆15,073Updated 2 weeks ago
- Text-To-Speech, RAG, and LLMs. All local!☆1,808Updated 6 months ago
- pingcap/autoflow is a Graph RAG based and conversational knowledge base tool built with TiDB Serverless Vector Storage. Demo: https://tid…☆2,593Updated 3 weeks ago