ngxson / wllamaLinks
WebAssembly binding for llama.cpp - Enabling on-browser LLM inference
☆762Updated last month
Alternatives and similar repositories for wllama
Users that are interested in wllama are comparing it to the libraries listed below
Sorting:
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆271Updated 11 months ago
- A cross-platform browser ML framework.☆708Updated 7 months ago
- Stateful load balancer custom-tailored for llama.cpp 🏓🦙☆790Updated this week
- VS Code extension for LLM-assisted code/text completion☆835Updated last week
- 🕸️🦀 A WASM vector similarity search written in Rust☆981Updated last year
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆575Updated 4 months ago
- Chat with AI large language models running natively in your browser. Enjoy private, server-free, seamless AI conversations.☆784Updated 2 months ago
- Vercel and web-llm template to run wasm models directly in the browser.☆158Updated last year
- FastMLX is a high performance production ready API to host MLX models.☆311Updated 3 months ago
- EntityDB is an in-browser vector database wrapping indexedDB and Transformers.js over WebAssembly☆187Updated 2 months ago
- Apple MLX engine for LM Studio☆669Updated this week
- ☆260Updated this week
- Big & Small LLMs working together☆1,058Updated this week
- An extremely fast implementation of whisper optimized for Apple Silicon using MLX.☆741Updated last year
- On-device LLM Inference Powered by X-Bit Quantization☆255Updated last month
- Run Large-Language Models (LLMs) 🚀 directly in your browser!☆210Updated 10 months ago
- Pure C++ implementation of several models for real-time chatting on your computer (CPU & GPU)☆656Updated this week
- Vectra is a local vector database for Node.js with features similar to pinecone but built using local files.☆498Updated 2 months ago
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.☆600Updated 8 months ago
- MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.☆1,470Updated this week
- ☆908Updated 10 months ago
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.☆447Updated 5 months ago
- LLM-powered lossless compression tool☆283Updated 10 months ago
- A collection of 🤗 Transformers.js demos and example applications☆1,658Updated last month
- llama.cpp fork with additional SOTA quants and improved performance☆652Updated this week
- Local AI API Platform☆2,765Updated last week
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.☆328Updated 2 weeks ago
- Suno AI's Bark model in C/C++ for fast text-to-speech generation☆831Updated 7 months ago
- Gemma 2 optimized for your local machine.☆376Updated 11 months ago
- E2B Desktop Sandbox for LLMs. E2B Sandbox with desktop graphical environment that you can connect to any LLM for secure computer use.☆1,000Updated last week