evilsocket / cake
Distributed LLM and StableDiffusion inference for mobile, desktop and server.
☆2,825Updated 5 months ago
Alternatives and similar repositories for cake:
Users that are interested in cake are comparing it to the libraries listed below
- Blazingly fast LLM inference.☆5,369Updated this week
- Connect home devices into a powerful cluster to accelerate LLM inference. More devices means faster inference.☆1,993Updated last week
- g1: Using Llama-3.1 70b on Groq to create o1-like reasoning chains☆4,201Updated 2 months ago
- Together Mixture-Of-Agents (MoA) – 65.1% on AlpacaEval with OSS models☆2,711Updated 2 months ago
- Official inference framework for 1-bit LLMs☆12,851Updated last month
- SCUDA is a GPU over IP bridge allowing GPUs on remote machines to be attached to CPU-only machines.☆1,682Updated 3 weeks ago
- A Datacenter Scale Distributed Inference Serving Framework☆3,377Updated this week
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,316Updated this week
- A framework for serving and evaluating LLM routers - save LLM costs without compromising quality☆3,763Updated 7 months ago
- A MLX port of FLUX based on the Huggingface Diffusers implementation.☆1,287Updated last week
- Enchanted is iOS and macOS app for chatting with private self hosted language models such as Llama2, Mistral or Vicuna using Ollama.☆5,087Updated last week
- AI app store powered by 24/7 desktop history. open source | 100% local | dev friendly | 24/7 screen, mic recording☆12,792Updated this week
- High-speed Large Language Model Serving for Local Deployment☆8,169Updated last month
- Ingest, parse, and optimize any data format ➡️ from documents to multimedia ➡️ for enhanced compatibility with GenAI frameworks☆6,418Updated 4 months ago
- Run PyTorch LLMs locally on servers, desktop and mobile☆3,548Updated this week
- AirLLM 70B inference with single 4GB GPU☆5,753Updated 4 months ago
- Local AI API Platform☆2,579Updated this week
- Sky-T1: Train your own O1 preview model within $450☆3,149Updated this week
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.☆1,105Updated this week
- Efficient Triton Kernels for LLM Training☆4,743Updated this week
- 🔍 An LLM-based Multi-agent Framework of Web Search Engine (like Perplexity.ai Pro and SearchGPT)☆6,250Updated 2 months ago
- RAG (Retrieval Augmented Generation) Framework for building modular, open source applications for production by TrueFoundry☆3,976Updated last month
- AICI: Prompts as (Wasm) Programs☆2,008Updated 2 months ago
- CoreNet: A library for training deep neural networks☆6,999Updated 5 months ago
- tiny vision language model☆7,701Updated this week
- Use your locally running AI models to assist you in your web browsing☆6,083Updated this week
- The easiest tool for fine-tuning LLM models, synthetic data generation, and collaborating on datasets.☆3,264Updated this week
- Claude Engineer is an interactive command-line interface (CLI) that leverages the power of Anthropic's Claude-3.5-Sonnet model to assist …☆10,936Updated 3 months ago
- AIOS: AI Agent Operating System☆3,992Updated last week
- Distributed Training Over-The-Internet☆893Updated 3 months ago