Picovoice / picollmLinks
On-device LLM Inference Powered by X-Bit Quantization
☆245Updated this week
Alternatives and similar repositories for picollm
Users that are interested in picollm are comparing it to the libraries listed below
Sorting:
- On-device streaming text-to-speech engine powered by deep learning☆85Updated this week
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆267Updated 10 months ago
- Recipes for on-device voice AI and local LLM☆85Updated this week
- FastMLX is a high performance production ready API to host MLX models.☆307Updated 2 months ago
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆182Updated 4 months ago
- Open source repo for AI in a Box.☆63Updated last year
- 1.58 Bit LLM on Apple Silicon using MLX☆213Updated last year
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆567Updated 3 months ago
- Start a server from the MLX library.☆187Updated 10 months ago
- Fast parallel LLM inference for MLX☆189Updated 11 months ago
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆737Updated last month
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆43Updated 8 months ago
- Local voice chatbot for engaging conversations, powered by Ollama, Hugging Face Transformers, and Coqui TTS Toolkit☆771Updated 9 months ago
- Open source LLM UI, compatible with all local LLM providers.☆174Updated 8 months ago
- Implementation of F5-TTS in MLX☆542Updated 2 months ago
- Pybind11 bindings for Whisper.cpp☆57Updated this week
- Awesome Mobile LLMs☆199Updated this week
- In-browser LLM website generator☆49Updated 4 months ago
- ML-powered speech synthesis directly in your browser☆156Updated 3 months ago
- ☆204Updated last year
- Big & Small LLMs working together☆920Updated this week
- A mobile Implementation of llama.cpp☆311Updated last year
- Blazing fast whisper turbo for ASR (speech-to-text) tasks☆208Updated 7 months ago
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆404Updated this week
- Fast Streaming TTS with Orpheus + WebRTC (with FastRTC)☆288Updated last month
- Compare open-source local LLM inference projects by their metrics to assess popularity and activeness.☆575Updated 3 weeks ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆268Updated last week
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.☆594Updated 7 months ago
- Run LLMs in the Browser with MLC / WebLLM ✨☆131Updated 8 months ago
- Local ML voice chat using high-end models.☆167Updated last week