ngxson / wllama
WebAssembly binding for llama.cpp - Enabling in-browser LLM inference
☆441Updated this week
Related projects ⓘ
Alternatives and complementary repositories for wllama
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆215Updated 4 months ago
- A cross-platform browser ML framework.☆627Updated this week
- FastMLX is a high performance production ready API to host MLX models.☆220Updated this week
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆496Updated 3 months ago
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆203Updated last month
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆237Updated 2 months ago
- MLX-VLM is a package for running Vision LLMs locally on your Mac using MLX.☆505Updated this week
- TypeScript generator for llama.cpp Grammar directly from TypeScript interfaces☆131Updated 4 months ago
- Stateful load balancer custom-tailored for llama.cpp 🏓🦙☆568Updated this week
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆228Updated this week
- ☆723Updated 2 months ago
- Efficient visual programming for AI language models☆304Updated 2 months ago
- Fast parallel LLM inference for MLX☆149Updated 4 months ago
- A simple UI / Web / Frontend for MLX mlx-lm using Streamlit.☆227Updated last month
- Chat with AI large language models running natively in your browser. Enjoy private, server-free, seamless AI conversations.☆469Updated last week
- Replace OpenAI with Llama.cpp Automagically.☆289Updated 5 months ago
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.☆501Updated 3 weeks ago
- Distributed Inference for mlx LLm☆70Updated 3 months ago
- 👾🍎 Apple MLX engine for LM Studio☆217Updated this week
- Implementation of F5-TTS in MLX☆332Updated 3 weeks ago
- LLM-powered lossless compression tool☆252Updated 3 months ago
- A fast, light, open chat UI with full tool use support across many models☆198Updated last month
- ☆229Updated last month
- SemanticFinder - frontend-only live semantic search with transformers.js☆233Updated 2 months ago
- ☆205Updated 2 months ago
- Start a server from the MLX library.☆161Updated 3 months ago
- 🚀 Retrieval Augmented Generation (RAG) with txtai. Combine search and LLMs to find insights with your own data.☆282Updated this week
- Gemma 2 optimized for your local machine.☆344Updated 3 months ago
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆110Updated 6 months ago
- ☆128Updated this week