hrishioa / wasm-ai
Vercel and web-llm template to run wasm models directly in the browser.
☆113Updated 9 months ago
Related projects: ⓘ
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆102Updated last year
- WebAssembly binding for llama.cpp - Enabling in-browser LLM inference☆342Updated last week
- Infrastructure powering E2B - Secure Runtime for AI Agents & Apps☆177Updated this week
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆183Updated last month
- SemanticFinder - frontend-only live semantic search with transformers.js☆212Updated last week
- 🦉 Open Source clone of Claude.ai that can generate Artifacts☆72Updated last month
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆126Updated 2 months ago
- Simple repo that compiles and runs llama2.c on the Web☆52Updated 10 months ago
- Web-optimized vector database (written in Rust).☆186Updated 2 months ago
- Framework to evaluate LLM generated ReactJS code.☆49Updated 5 months ago
- A browser-based AI inference network.☆112Updated last month
- A client side vector search library that can embed, store, search, and cache vectors. Works on the browser and node. It outperforms OpenA…☆156Updated 3 months ago
- GPU accelerated client-side embeddings for vector search, RAG etc.☆63Updated 9 months ago
- ☆133Updated 9 months ago
- A collection of LLM services you can self host via docker or modal labs to support your applications development☆181Updated 4 months ago
- Demo of AI chatbot that predicts user message to generate response quickly.☆95Updated 6 months ago
- An HTTP serving framework by Banana☆97Updated 9 months ago
- ☆77Updated this week
- WebGPU LLM inference tuned by hand☆145Updated last year
- Run GGML models with Kubernetes.☆172Updated 9 months ago
- The easiest, and fastest way to run AI-generated Python code safely☆188Updated 3 months ago
- 🌸 The open framework for question answering fine-tuning LLMs on private data☆69Updated 10 months ago
- 🤖 Headless IDE for AI agents☆110Updated this week
- ☆70Updated this week
- Structured inference with Llama 2 in your browser☆48Updated 3 months ago
- JS tokenizer for LLaMA 1 and 2☆330Updated 2 months ago
- Vector Storage is a vector database that enables semantic similarity searches on text documents in the browser's local storage. It uses O…☆190Updated 10 months ago
- Edge full-stack LLM platform. Written in Rust☆369Updated 3 months ago
- Natural Language Interfaces Powered by LLMs☆91Updated last month
- Action library for AI Agent☆187Updated this week