LlamaEdge / sd-api-serverLinks
The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge
☆24Updated 11 months ago
Alternatives and similar repositories for sd-api-server
Users that are interested in sd-api-server are comparing it to the libraries listed below
Sorting:
- A RAG API server written in Rust following OpenAI specs☆60Updated 9 months ago
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆27Updated 10 months ago
- A Pure Rust based LLM, VLM, VLA, TTS, OCR Inference Engine, powering by Candle & Rust. Alternate to your llama.cpp but much more simpler …☆241Updated last week
- The MCP enterprise actors-based server or mcp-ectors for short☆31Updated 8 months ago
- The Google mediapipe AI library. Write AI inference applications for image recognition, text classification, audio / video processing and…☆227Updated 3 weeks ago
- Blazingly fast inference of diffusion models.☆119Updated 10 months ago
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆47Updated last year
- An educational Rust project for exporting and running inference on Qwen3 LLM family☆40Updated 6 months ago
- 🦀 A Pure Rust Framework For Building AGI (WIP).☆111Updated 3 weeks ago
- 🔥🔥 Kokoro in Rust. https://huggingface.co/hexgrad/Kokoro-82M Insanely fast, realtime TTS with high quality you ever have.☆701Updated 2 weeks ago
- AI gateway and observability server written in Rust. Designed to help optimize multi-agent workflows.☆65Updated last year
- Rust bindings to https://github.com/k2-fsa/sherpa-onnx☆287Updated 3 months ago
- Simple Rust applications that run in WasmEdge☆33Updated 2 years ago
- AI Assistant☆20Updated 9 months ago
- A Fish Speech implementation in Rust, with Candle.rs☆107Updated 8 months ago
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes☆242Updated 6 months ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated 2 years ago
- Embed WasmEdge functions in a Rust host app☆33Updated last year
- An LLM interface (chat bot) implemented in pure Rust using HuggingFace/Candle over Axum Websockets, an SQLite Database, and a Leptos (Was…☆139Updated last year
- Lightweight web service clients in the WasmEdge Runtime using the Rust reqwest framework☆12Updated last year
- Low rank adaptation (LoRA) for Candle.☆169Updated 9 months ago
- ☆13Updated 9 months ago
- Rust bindings for OpenNMT/CTranslate2☆49Updated last month
- Built for demanding AI workflows, this gateway offers low-latency, provider-agnostic access, ensuring your AI applications run smoothly a…☆89Updated 8 months ago
- Kokoro TTS的Rust推理实现☆29Updated 2 weeks ago
- LM inference server implementation based on *.cpp.☆295Updated 2 months ago
- Fast serverless LLM inference, in Rust.☆109Updated 3 months ago
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆266Updated 11 months ago
- Implementation of the RWKV language model in pure WebGPU/Rust.☆338Updated 3 weeks ago
- Minimalistic Rust Implementation Of Model Context Protocol from Anthropic☆63Updated 6 months ago