second-state / WasmEdge-WASINN-examples
☆245Updated this week
Alternatives and similar repositories for WasmEdge-WASINN-examples:
Users that are interested in WasmEdge-WASINN-examples are comparing it to the libraries listed below
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,260Updated this week
- Neural Network proposal for WASI☆490Updated 4 months ago
- OpenAI compatible API for serving LLAMA-2 model☆215Updated last year
- The Google mediapipe AI library. Write AI inference applications for image recognition, text classification, audio / video processing and…☆172Updated 5 months ago
- A cross-platform browser ML framework.☆666Updated 3 months ago
- Vercel and web-llm template to run wasm models directly in the browser.☆140Updated last year
- Moly: an AI LLM GUI app in pure Rust☆202Updated this week
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆320Updated this week
- LLM Orchestrator built in Rust☆272Updated 11 months ago
- A WebGPU-accelerated ONNX inference run-time written 100% in Rust, ready for native and the web☆1,707Updated 7 months ago
- Tensor library for machine learning☆278Updated last year
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes☆183Updated 2 weeks ago
- LLama.cpp rust bindings☆374Updated 8 months ago
- Run any ML model from any programming language.☆422Updated last year
- a fast cross platform AI inference engine 🤖 using Rust 🦀 and WebGPU 🎮☆432Updated 2 months ago
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆102Updated last year
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆604Updated this week
- A RAG API server written in Rust following OpenAI specs☆44Updated last week
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- A template project to demonstrate how to run WebAssembly functions as sidecar microservices in dapr☆279Updated last year
- Inference Llama 2 in one file of pure Rust 🦀☆232Updated last year
- Generate ideal question-answers for testing RAG☆126Updated last week
- ☆224Updated this week
- Simple repo that compiles and runs llama2.c on the Web☆54Updated last year
- ☆135Updated last year
- Rust framework for LLM orchestration☆202Updated 7 months ago
- Implementation of the RWKV language model in pure WebGPU/Rust.☆288Updated this week
- High-level, optionally asynchronous Rust bindings to llama.cpp☆208Updated 9 months ago
- Minimal LLM inference in Rust☆975Updated 4 months ago
- Rust+OpenCL+AVX2 implementation of LLaMA inference code☆542Updated last year