ngxson / wllamaLinks
WebAssembly binding for llama.cpp - Enabling on-browser LLM inference
☆843Updated 3 weeks ago
Alternatives and similar repositories for wllama
Users that are interested in wllama are comparing it to the libraries listed below
Sorting:
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆275Updated last year
- VS Code extension for LLM-assisted code/text completion☆917Updated this week
- ☆306Updated this week
- Open-source LLMOps platform for hosting and scaling AI in your own infrastructure 🏓🦙☆1,098Updated this week
- LM Studio Apple MLX engine☆745Updated 2 weeks ago
- 🕸️🦀 A WASM vector similarity search written in Rust☆994Updated last year
- FastMLX is a high performance production ready API to host MLX models.☆324Updated 5 months ago
- Big & Small LLMs working together☆1,127Updated this week
- Vercel and web-llm template to run wasm models directly in the browser.☆160Updated last year
- Vectra is a local vector database for Node.js with features similar to pinecone but built using local files.☆512Updated 3 months ago
- Chat with AI large language models running natively in your browser. Enjoy private, server-free, seamless AI conversations.☆824Updated 3 months ago
- EntityDB is an in-browser vector database wrapping indexedDB and Transformers.js over WebAssembly☆204Updated 3 months ago
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆587Updated 6 months ago
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.☆1,583Updated this week
- JS tokenizer for LLaMA 3 and LLaMA 3.1☆116Updated 3 weeks ago
- On-device LLM Inference Powered by X-Bit Quantization☆266Updated 2 weeks ago
- Run Large-Language Models (LLMs) 🚀 directly in your browser!☆214Updated 11 months ago
- An extremely fast implementation of whisper optimized for Apple Silicon using MLX.☆769Updated last year
- Suno AI's Bark model in C/C++ for fast text-to-speech generation☆834Updated 9 months ago
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.☆453Updated 6 months ago
- ML-powered speech synthesis directly in your browser☆165Updated 6 months ago
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆140Updated last year
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆531Updated this week
- Gemma 2 optimized for your local machine.☆377Updated last year
- SemanticFinder - frontend-only live semantic search with transformers.js☆292Updated 4 months ago
- LLM-powered lossless compression tool☆288Updated last year
- The easiest & fastest way to run customized and fine-tuned LLMs locally or on the edge☆1,476Updated this week
- Local AI API Platform☆2,762Updated last month
- A SQLite extension for generating text embeddings from GGUF models using llama.cpp☆211Updated 9 months ago
- Official inference library for pre-processing of Mistral models☆778Updated 2 weeks ago