Picovoice / picollmLinks
On-device LLM Inference Powered by X-Bit Quantization
☆269Updated 2 months ago
Alternatives and similar repositories for picollm
Users that are interested in picollm are comparing it to the libraries listed below
Sorting:
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 8 months ago
- Recipes for on-device voice AI and local LLM☆99Updated 4 months ago
- On-device streaming text-to-speech engine powered by deep learning☆122Updated last month
- ☆413Updated last week
- 1.58 Bit LLM on Apple Silicon using MLX☆224Updated last year
- Open source LLM UI, compatible with all local LLM providers.☆175Updated last year
- A platform to self-host AI on easy mode☆170Updated this week
- Running a LLM on the ESP32☆78Updated last year
- Pybind11 bindings for Whisper.cpp☆62Updated 2 weeks ago
- Run LLMs in the Browser with MLC / WebLLM ✨☆140Updated last year
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆273Updated last year
- API Server for Transformer Lab☆79Updated last week
- Locally running LLM with internet access☆97Updated 3 months ago
- Replace OpenAI with Llama.cpp Automagically.☆324Updated last year
- Fast Streaming TTS with Orpheus + WebRTC (with FastRTC)☆336Updated 6 months ago
- plug whisper audio transcription to a local ollama server and ouput tts audio responses☆351Updated last year
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆278Updated 4 months ago
- ☆132Updated 5 months ago
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS over OpenAI endpoints.☆213Updated last week
- Local ML voice chat using high-end models.☆175Updated last month
- A memory framework for Large Language Models and Agents.☆183Updated 9 months ago
- ☆91Updated 5 months ago
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆282Updated last year
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆599Updated 8 months ago
- ☆207Updated last year
- Code for Papeg.ai☆224Updated 9 months ago
- FastMLX is a high performance production ready API to host MLX models.☆331Updated 7 months ago
- A fast batching API to serve LLM models☆188Updated last year
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆261Updated 7 months ago
- A simple experiment on letting two local LLM have a conversation about anything!☆111Updated last year