ngxson / wllamaLinks
WebAssembly binding for llama.cpp - Enabling on-browser LLM inference
β750Updated 2 weeks ago
Alternatives and similar repositories for wllama
Users that are interested in wllama are comparing it to the libraries listed below
Sorting:
- WebAssembly (Wasm) Build and Bindings for llama.cppβ269Updated 10 months ago
- Stateful load balancer custom-tailored for llama.cpp ππ¦β779Updated this week
- A cross-platform browser ML framework.β702Updated 6 months ago
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM β¦β573Updated 4 months ago
- FastMLX is a high performance production ready API to host MLX models.β308Updated 3 months ago
- VS Code extension for LLM-assisted code/text completionβ807Updated this week
- EntityDB is an in-browser vector database wrapping indexedDB and Transformers.js over WebAssemblyβ177Updated last month
- Apple MLX engine for LM Studioβ618Updated last month
- πΈοΈπ¦ A WASM vector similarity search written in Rustβ976Updated last year
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. Iβ¦β418Updated last week
- llama.cpp fork with additional SOTA quants and improved performanceβ584Updated this week
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.β597Updated 7 months ago
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.β446Updated 4 months ago
- An extremely fast implementation of whisper optimized for Apple Silicon using MLX.β724Updated last year
- Model swapping for llama.cpp (or any local OpenAPI compatible server)β961Updated this week
- Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)β627Updated this week
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edgeβ1,431Updated last week
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.β1,347Updated 2 weeks ago
- Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation leβ¦β1,549Updated last week
- A multi-platform desktop application to evaluate and compare LLM models, written in Rust and React.β757Updated 2 months ago
- β900Updated 9 months ago
- Big & Small LLMs working togetherβ972Updated this week
- A collection of π€ Transformers.js demos and example applicationsβ1,603Updated 2 weeks ago
- Local AI API Platformβ2,753Updated last week
- Chat with AI large language models running natively in your browser. Enjoy private, server-free, seamless AI conversations.β769Updated last month
- Sidecar is the AI brains for the Aide editor and works alongside it, locally on your machineβ570Updated last month
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfacesβ137Updated 11 months ago
- LLM-based code completion engineβ194Updated 4 months ago
- Vercel and web-llm template to run wasm models directly in the browser.β152Updated last year
- On-device LLM Inference Powered by X-Bit Quantizationβ249Updated last week