ngxson / wllama
WebAssembly binding for llama.cpp - Enabling on-browser LLM inference
β582Updated last week
Alternatives and similar repositories for wllama:
Users that are interested in wllama are comparing it to the libraries listed below
- WebAssembly (Wasm) Build and Bindings for llama.cppβ235Updated 6 months ago
- A cross-platform browser ML framework.β658Updated 2 months ago
- Stateful load balancer custom-tailored for llama.cpp ππ¦β706Updated last month
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM β¦β530Updated this week
- VS Code extension for LLM-assisted code/text completionβ532Updated this week
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.β845Updated 2 weeks ago
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfacesβ134Updated 7 months ago
- 1.58 Bit LLM on Apple Silicon using MLXβ184Updated 9 months ago
- An extremely fast implementation of whisper optimized for Apple Silicon using MLX.β646Updated 9 months ago
- FastMLX is a high performance production ready API to host MLX models.β260Updated 2 months ago
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I β¦β255Updated this week
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.β541Updated 3 months ago
- function calling-based LLM agentsβ283Updated 5 months ago
- πΎπ Apple MLX engine for LM Studioβ396Updated this week
- Fast parallel LLM inference for MLXβ163Updated 7 months ago
- Sidecar is the AI brains for the Aide editor and works alongside it, locally on your machineβ476Updated this week
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.β252Updated this week
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.β417Updated 3 weeks ago
- Vercel and web-llm template to run wasm models directly in the browser.β137Updated last year
- LLM-based code completion engineβ179Updated 3 weeks ago
- Infrastructure for AI code interpreting that's powering E2B.β289Updated this week
- A simple UI / Web / Frontend for MLX mlx-lm using Streamlit.β242Updated 3 weeks ago
- πΈοΈπ¦ A WASM vector similarity search written in Rustβ919Updated last year
- A multi-platform desktop application to evaluate and compare LLM models, written in Rust and React.β642Updated last month
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Siliconβ260Updated 5 months ago
- Start a server from the MLX library.β173Updated 6 months ago
- Chat with AI large language models running natively in your browser. Enjoy private, server-free, seamless AI conversations.β624Updated 3 weeks ago
- Simple Python library/structure to ablate features in LLMs which are supported by TransformerLensβ412Updated 8 months ago
- EntityDB is an in-browser vector database wrapping indexedDB and Transformers.js over WebAssemblyβ116Updated last month
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edgeβ1,249Updated this week