Picovoice / picollmLinks
On-device LLM Inference Powered by X-Bit Quantization
☆273Updated last week
Alternatives and similar repositories for picollm
Users that are interested in picollm are comparing it to the libraries listed below
Sorting:
- On-device streaming text-to-speech engine powered by deep learning☆121Updated last week
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 10 months ago
- ☆547Updated this week
- Local ML voice chat using high-end models.☆178Updated last week
- Recipes for on-device voice AI and local LLM☆100Updated last month
- Open source LLM UI, compatible with all local LLM providers.☆176Updated last year
- Run LLMs in the Browser with MLC / WebLLM ✨☆150Updated last year
- A platform to self-host AI on easy mode☆181Updated last week
- 1.58 Bit LLM on Apple Silicon using MLX☆227Updated last year
- Vercel and web-llm template to run wasm models directly in the browser.☆166Updated 2 years ago
- WebAssembly (Wasm) Build and Bindings for llama.cpp☆285Updated last year
- Using FastChat-T5 Large Language Model, Vosk API for automatic speech recognition, and Piper for text-to-speech☆126Updated 2 years ago
- No-code CLI designed for accelerating ONNX workflows☆221Updated 6 months ago
- Something similar to Apple Intelligence?☆61Updated last year
- ☆210Updated 3 months ago
- A fully in-browser privacy solution to make Conversational AI privacy-friendly☆233Updated last year
- Blazing fast whisper turbo for ASR (speech-to-text) tasks☆217Updated last month
- Running a LLM on the ESP32☆83Updated last year
- In-browser LLM website generator☆50Updated 10 months ago
- A simple experiment on letting two local LLM have a conversation about anything!☆112Updated last year
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆285Updated 6 months ago
- Fast parallel LLM inference for MLX☆235Updated last year
- Setup and run a local LLM and Chatbot using consumer grade hardware.☆303Updated 3 weeks ago
- Distributed Inference for mlx LLm☆99Updated last year
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆266Updated 9 months ago
- Enhancing LLMs with LoRA☆193Updated 2 months ago
- ☆108Updated 4 months ago
- ☆134Updated last week
- A fast batching API to serve LLM models☆189Updated last year
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆261Updated last week