Picovoice / picollmLinks
On-device LLM Inference Powered by X-Bit Quantization
☆274Updated this week
Alternatives and similar repositories for picollm
Users that are interested in picollm are comparing it to the libraries listed below
Sorting:
- On-device streaming text-to-speech engine powered by deep learning☆127Updated last week
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 11 months ago
- Recipes for on-device voice AI and local LLM☆103Updated this week
- Run LLMs in the Browser with MLC / WebLLM ✨☆150Updated last year
- Aggregates compute from spare GPU capacity☆183Updated last week
- 1.58 Bit LLM on Apple Silicon using MLX☆237Updated last year
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆118Updated last year
- Open source LLM UI, compatible with all local LLM providers.☆177Updated last year
- ☆134Updated last month
- Something similar to Apple Intelligence?☆59Updated last year
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆274Updated 2 months ago
- A simple experiment on letting two local LLM have a conversation about anything!☆112Updated last year
- FastMLX is a high performance production ready API to host MLX models.☆339Updated 9 months ago
- Locally running LLM with internet access☆97Updated 6 months ago
- Local ML voice chat using high-end models.☆178Updated 3 weeks ago
- Awesome Mobile LLMs☆290Updated last month
- Setup and run a local LLM and Chatbot using consumer grade hardware.☆308Updated last month
- Replace OpenAI with Llama.cpp Automagically.☆326Updated last year
- API Server for Transformer Lab☆82Updated last month
- This small API downloads and exposes access to NeuML's txtai-wikipedia and full wikipedia datasets, taking in a query and returning full …☆101Updated 4 months ago
- An application for running LLMs locally on your device, with your documents, facilitating detailed citations in generated responses.☆624Updated last year
- Running a LLM on the ESP32☆85Updated last year
- Fast Streaming TTS with Orpheus + WebRTC (with FastRTC)☆347Updated 9 months ago
- Code for Papeg.ai☆226Updated last year
- ☆210Updated last week
- LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.☆130Updated 2 years ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆286Updated 6 months ago
- ☆94Updated last year
- A memory framework for Large Language Models and Agents.☆182Updated last year
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆99Updated 6 months ago