huggingface / ratchet
A cross-platform browser ML framework.
β567Updated last week
Related projects: β
- WebAssembly binding for llama.cpp - Enabling in-browser LLM inferenceβ342Updated 2 weeks ago
- πΈοΈπ¦ A WASM vector similarity search written in Rustβ868Updated last year
- LLM Orchestrator built in Rustβ261Updated 6 months ago
- β134Updated 7 months ago
- LLama.cpp rust bindingsβ320Updated 2 months ago
- High-level, optionally asynchronous Rust bindings to llama.cppβ161Updated 3 months ago
- Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B.β373Updated 6 months ago
- β2,155Updated this week
- β134Updated this week
- Stateful load balancer custom-tailored for llama.cppβ523Updated this week
- On-device Inference of Diffusion Models for Apple Siliconβ437Updated last week
- Library for generating vector embeddings, reranking in Rustβ250Updated 3 weeks ago
- A Rust implementation of OpenAI's Whisper model using the burn frameworkβ262Updated 4 months ago
- An extremely fast implementation of whisper optimized for Apple Silicon using MLX.β522Updated 4 months ago
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.β229Updated 3 weeks ago
- Web-optimized vector database (written in Rust).β186Updated 2 months ago
- π¦ A curated list of Rust tools, libraries, and frameworks for working with LLMs, GPT, AIβ236Updated 6 months ago
- β122Updated 4 months ago
- OpenAI compatible API for serving LLAMA-2 modelβ212Updated 11 months ago
- β232Updated 3 weeks ago
- WebAssembly (Wasm) Build and Bindings for llama.cppβ183Updated last month
- A multi-platform desktop application to evaluate and compare LLM models, written in Rust and React.β430Updated this week
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Siliconβ206Updated last week
- Whisper with Medusa headsβ774Updated last week
- Inference Llama 2 in one file of pure Rust π¦β227Updated last year
- A WebGPU-accelerated ONNX inference run-time written 100% in Rust, ready for native and the webβ1,605Updated last month
- A minimalist yet highly performant, lightweight, lightning fast, multisource, multimodal and local embedding solution, built in rust.β220Updated this week
- FastMLX is a high performance production ready API to host MLX models.β163Updated last week
- β617Updated this week
- Tutorial for Porting PyTorch Transformer Models to Candle (Rust)β235Updated last month