A Pure Rust based LLM, VLM, VLA, TTS, OCR Inference Engine, powering by Candle & Rust. Alternate to your llama.cpp but much more simpler and cleaner..
β351Apr 15, 2026Updated this week
Alternatives and similar repositories for Crane
Users that are interested in Crane are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Rust standalone inference of Namo-500M series models. Extremly tiny, runing VLM on CPU.β24Mar 12, 2025Updated last year
- π₯π₯ Kokoro in Rust. https://huggingface.co/hexgrad/Kokoro-82M Insanely fast, realtime TTS with high quality you ever have.β754Mar 11, 2026Updated last month
- β24Jan 22, 2025Updated last year
- Cleanai (https://github.com/willmil11/cleanai) except I'm making it in c now. Fast and clean from the start this time :)β17Mar 6, 2026Updated last month
- A Rust π¦ port of the Hugging Face smolagents library.β42Mar 26, 2025Updated last year
- 1-Click AI Models by DigitalOcean Gradient β’ AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- Game servers running on Kubernetesβ12Apr 28, 2025Updated 11 months ago
- Implementation of Qwen3-ASR-0.6B in GGMLβ77Feb 10, 2026Updated 2 months ago
- Run Orpheus 3B Locally with Gradio UI, Standalone Appβ24Apr 1, 2025Updated last year
- An easy-to-use library and command-line tool for TTSβ15May 3, 2025Updated 11 months ago
- Adding a multi-text multi-speaker script (diffe) that is based on a script from asiff00 on issue 61 for Sesame: A Conversational Speech Gβ¦β26Mar 28, 2025Updated last year
- Rust bindings to https://github.com/k2-fsa/sherpa-onnxβ307Mar 8, 2026Updated last month
- Fast, flexible LLM inferenceβ6,994Updated this week
- β535Updated this week
- The Python Implementation of CRISP: Clustering Multi-Vector Representations for Denoising and Pruningβ27Jul 27, 2025Updated 8 months ago
- Simple, predictable pricing with DigitalOcean hosting β’ AdAlways know what you'll pay with monthly caps and flat pricing. Enterprise-grade infrastructure trusted by 600k+ customers.
- β52Feb 19, 2025Updated last year
- the rent a hal project for AIβ21Apr 11, 2026Updated last week
- A Rust-based, SenseVoiceSmallβ31Updated this week
- The hearth of The Pulsar App, fast, secure and shared inference with modern UIβ60Dec 1, 2024Updated last year
- superfast text to speech in any voiceβ62Feb 16, 2026Updated 2 months ago
- β16May 14, 2025Updated 11 months ago
- β15Mar 18, 2026Updated last month
- β65Jun 24, 2025Updated 9 months ago
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Faceβ48May 3, 2024Updated last year
- Wordpress hosting with auto-scaling - Free Trial β’ AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- LLM inference in C/C++β23Oct 4, 2024Updated last year
- A forward proxy to turn network traffic into personal memory for AI agentsβ38Mar 30, 2026Updated 2 weeks ago
- Service for testing out the new Qwen2.5 omni modelβ63Apr 30, 2025Updated 11 months ago
- An educational Rust project for exporting and running inference on Qwen3 LLM familyβ42Aug 3, 2025Updated 8 months ago
- High-performance Qwen3-TTS implementation | Instruction-driven Β· Zero-shot voice cloning Β· Streaming Β· RTF 0.55β56Apr 4, 2026Updated 2 weeks ago
- A pure and fast NumPy implementation of Mamba with cache support.β18Jun 16, 2024Updated last year
- Scripts and tools for optimizing quantizations in llama.cpp with GGUF imatrices.β19Jan 10, 2025Updated last year
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.β641Updated this week
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibesβ248Aug 6, 2025Updated 8 months ago
- GPU virtual machines on DigitalOcean Gradient AI β’ AdGet to production fast with high-performance AMD and NVIDIA GPUs you can spin up in seconds. The definition of operational simplicity.
- Fast ML inference & training for ONNX models in Rustβ2,181Updated this week
- β18Aug 19, 2025Updated 8 months ago
- A collection of experimental Retrieval Augmented Generation (RAG) Techniques to elevate your pipelines, all with code and intuitive explaβ¦β36Jul 21, 2025Updated 8 months ago
- AI Assistantβ20Feb 21, 2026Updated last month
- High-level, optionally asynchronous Rust bindings to llama.cppβ245Jun 5, 2024Updated last year
- β49Mar 17, 2025Updated last year
- β20Jul 4, 2025Updated 9 months ago