huggingface / ratchet
A cross-platform browser ML framework.
β652Updated last month
Alternatives and similar repositories for ratchet:
Users that are interested in ratchet are comparing it to the libraries listed below
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inferenceβ485Updated this week
- πΈοΈπ¦ A WASM vector similarity search written in Rustβ902Updated last year
- LLM Orchestrator built in Rustβ267Updated 10 months ago
- LLama.cpp rust bindingsβ354Updated 6 months ago
- Rust library for generating vector embeddings, reranking locallyβ397Updated this week
- A Rust implementation of OpenAI's Whisper model using the burn frameworkβ284Updated 8 months ago
- β136Updated 10 months ago
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.β289Updated this week
- High-level, optionally asynchronous Rust bindings to llama.cppβ194Updated 7 months ago
- β193Updated this week
- Ready-made tokenizer library for working with GPT and tiktokenβ277Updated last week
- Fast, streaming indexing, query, and agent library for building LLM applications in Rustβ348Updated this week
- Web-optimized vector database (written in Rust).β198Updated this week
- Rust client for the huggingface hub aiming for minimal subset of features over `huggingface-hub` python packageβ169Updated this week
- Rust+OpenCL+AVX2 implementation of LLaMA inference codeβ541Updated 11 months ago
- Tutorial for Porting PyTorch Transformer Models to Candle (Rust)β272Updated 5 months ago
- Like grep but for natural language questions. Based on Mistral 7B or Mixtral 8x7B.β378Updated 10 months ago
- Stateful load balancer custom-tailored for llama.cpp ππ¦β666Updated last week
- Inference Llama 2 in one file of pure Rust π¦β231Updated last year
- An implementation of the diffusers api in Rustβ546Updated 9 months ago
- Production-Ready Inference, Ingestion and Indexing built in Rust π¦β387Updated this week
- A WebGPU-accelerated ONNX inference run-time written 100% in Rust, ready for native and the webβ1,693Updated 5 months ago
- β123Updated 8 months ago
- π¦ A curated list of Rust tools, libraries, and frameworks for working with LLMs, GPT, AIβ318Updated 10 months ago
- WebAssembly (Wasm) Build and Bindings for llama.cppβ225Updated 5 months ago
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Siliconβ255Updated 4 months ago
- An extremely fast implementation of whisper optimized for Apple Silicon using MLX.β615Updated 8 months ago
- Low rank adaptation (LoRA) for Candle.β134Updated 4 months ago
- Vercel and web-llm template to run wasm models directly in the browser.β133Updated last year
- Minimal LLM inference in Rustβ958Updated 2 months ago