tangledgroup / llama-cpp-wasmLinks
WebAssembly (Wasm) Build and Bindings for llama.cpp
☆267Updated 10 months ago
Alternatives and similar repositories for llama-cpp-wasm
Users that are interested in llama-cpp-wasm are comparing it to the libraries listed below
Sorting:
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆732Updated last month
- Vercel and web-llm template to run wasm models directly in the browser.☆150Updated last year
- Run Large-Language Models (LLMs) 🚀 directly in your browser!☆207Updated 8 months ago
- Inference Llama 2 in one file of pure JavaScript(HTML)☆33Updated 2 weeks ago
- A cross-platform browser ML framework.☆696Updated 6 months ago
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆137Updated 10 months ago
- A JavaScript library that brings vector search and RAG to your browser!☆119Updated 9 months ago
- EntityDB is an in-browser vector database wrapping indexedDB and Transformers.js over WebAssembly☆168Updated 3 weeks ago
- JS tokenizer for LLaMA 3 and LLaMA 3.1☆108Updated 2 months ago
- JS tokenizer for LLaMA 1 and 2☆350Updated 11 months ago
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆566Updated 3 months ago
- Implementation of the RWKV language model in pure WebGPU/Rust.☆307Updated 2 weeks ago
- LLM-based code completion engine☆189Updated 4 months ago
- Browser-compatible JS library for running language models☆228Updated 2 years ago
- Add local LLMs to your Web or Electron apps! Powered by Rust + WebGPU☆102Updated last year
- Tensor library for machine learning☆275Updated 2 years ago
- On-device LLM Inference Powered by X-Bit Quantization☆241Updated 3 weeks ago
- A simple vector database built on idb☆85Updated last year
- A client side vector search library that can embed, store, search, and cache vectors. Works on the browser and node. It outperforms OpenA…☆203Updated last year
- SemanticFinder - frontend-only live semantic search with transformers.js☆273Updated 2 months ago
- WebGPU LLM inference tuned by hand☆150Updated last year
- Vectra is a local vector database for Node.js with features similar to pinecone but built using local files.☆480Updated 3 weeks ago
- Library to generate vector embeddings in NodeJS☆121Updated last month
- ggml implementation of BERT☆491Updated last year
- Fast parallel LLM inference for MLX☆189Updated 10 months ago
- Distributed Inference for mlx LLm☆92Updated 10 months ago
- Stateful load balancer custom-tailored for llama.cpp 🏓🦙☆764Updated last week
- Web-optimized vector database (written in Rust).☆238Updated 3 months ago
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.☆318Updated 2 weeks ago
- Python bindings for ggml☆140Updated 9 months ago