second-state / WasmEdge-WASINN-examples
☆232Updated 3 weeks ago
Related projects: ⓘ
- A cross-platform browser ML framework.☆567Updated last week
- OpenAI compatible API for serving LLAMA-2 model☆212Updated 11 months ago
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆990Updated this week
- WebAssembly binding for llama.cpp - Enabling in-browser LLM inference☆342Updated 2 weeks ago
- The Google mediapipe AI library. Write AI inference applications for image recognition, text classification, audio / video processing and…☆141Updated last month
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆229Updated 3 weeks ago
- a fast cross platform AI inference engine 🤖 using Rust 🦀 and WebGPU 🎮☆394Updated last month
- Vercel and web-llm template to run wasm models directly in the browser.☆113Updated 10 months ago
- A WebGPU-accelerated ONNX inference run-time written 100% in Rust, ready for native and the web☆1,605Updated last month
- Run any ML model from any programming language.☆421Updated 8 months ago
- ☆134Updated 7 months ago
- Approx nearest neighbor search in Rust☆160Updated last year
- 🦀 A curated list of Rust tools, libraries, and frameworks for working with LLMs, GPT, AI☆236Updated 6 months ago
- LLM Orchestrator built in Rust☆261Updated 6 months ago
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆102Updated last year
- LLama.cpp rust bindings☆320Updated 2 months ago
- 🦀Rust + Large Language Models - Make AI Services Freely and Easily.☆180Updated 6 months ago
- Inference Llama 2 in one file of pure Rust 🦀☆227Updated last year
- Web-optimized vector database (written in Rust).☆186Updated 2 months ago
- Tutorial for Porting PyTorch Transformer Models to Candle (Rust)☆235Updated last month
- Tensor library for machine learning☆278Updated last year
- Hybrid vector database with flexible SQL storage engine & multi-index support.☆343Updated this week
- ☆134Updated this week
- JS tokenizer for LLaMA 1 and 2☆330Updated 2 months ago
- WebGPU LLM inference tuned by hand☆145Updated last year
- A fast llama2 decoder in pure Rust.☆1,005Updated 9 months ago
- High-level, optionally asynchronous Rust bindings to llama.cpp☆161Updated 3 months ago
- ☆190Updated last week
- Implementation of the RWKV language model in pure WebGPU/Rust.☆227Updated last month
- Fast, streaming indexing and query library for AI (RAG) applications, written in Rust☆129Updated this week