tangledgroup / llama-cpp-wasm
WebAssembly (Wasm) Build and Bindings for llama.cpp
☆235Updated 6 months ago
Alternatives and similar repositories for llama-cpp-wasm:
Users that are interested in llama-cpp-wasm are comparing it to the libraries listed below
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆582Updated last week
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆134Updated 7 months ago
- Vercel and web-llm template to run wasm models directly in the browser.☆137Updated last year
- A cross-platform browser ML framework.☆658Updated 2 months ago
- Run Large-Language Models (LLMs) 🚀 directly in your browser!☆180Updated 5 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆184Updated 9 months ago
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆102Updated last year
- WebGPU LLM inference tuned by hand☆148Updated last year
- JS tokenizer for LLaMA 3 and LLaMA 3.1☆103Updated 6 months ago
- Vector Storage is a vector database that enables semantic similarity searches on text documents in the browser's local storage. It uses O…☆214Updated 2 months ago
- Simple repo that compiles and runs llama2.c on the Web☆54Updated last year
- Use self-hosted LLMs with an OpenAI compatible API☆63Updated 10 months ago
- A fully in-browser privacy solution to make Conversational AI privacy-friendly☆227Updated 4 months ago
- A JavaScript implementation of Llama 3 using node-mlx.☆72Updated 7 months ago
- Simple LLM library for JavaScript☆38Updated 3 weeks ago
- JavaScript implementation of LiteLLM.☆113Updated last month
- Web-optimized vector database (written in Rust).☆208Updated last week
- ☆31Updated last year
- JS tokenizer for LLaMA 1 and 2☆349Updated 7 months ago
- Vectra is a local vector database for Node.js with features similar to pinecone but built using local files.☆436Updated 3 months ago
- Fast parallel LLM inference for MLX☆163Updated 7 months ago
- A client side vector search library that can embed, store, search, and cache vectors. Works on the browser and node. It outperforms OpenA…☆185Updated 8 months ago
- LLM-based code completion engine☆179Updated 3 weeks ago
- ☆35Updated last year
- ☆228Updated 3 months ago
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆75Updated last year
- llama.cpp fork with additional SOTA quants and improved performance☆155Updated this week
- Port of Suno AI's Bark in C/C++ for fast inference☆55Updated 10 months ago
- Infrastructure for AI code interpreting that's powering E2B.☆289Updated this week
- A multimodal, function calling powered LLM webui.☆214Updated 4 months ago