Picovoice / picollmLinks
On-device LLM Inference Powered by X-Bit Quantization
☆249Updated 2 weeks ago
Alternatives and similar repositories for picollm
Users that are interested in picollm are comparing it to the libraries listed below
Sorting:
- On-device streaming text-to-speech engine powered by deep learning☆87Updated last week
- 1.58 Bit LLM on Apple Silicon using MLX☆214Updated last year
- Recipes for on-device voice AI and local LLM☆88Updated 2 weeks ago
- Running a LLM on the ESP32☆70Updated 8 months ago
- ☆234Updated this week
- Start a server from the MLX library.☆187Updated 11 months ago
- Using FastChat-T5 Large Language Model, Vosk API for automatic speech recognition, and Piper for text-to-speech☆120Updated 2 years ago
- Awesome Mobile LLMs☆204Updated 3 weeks ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆273Updated last week
- Pybind11 bindings for Whisper.cpp☆58Updated 3 weeks ago
- Lightweight Inference server for OpenVINO☆187Updated last week
- Open source LLM UI, compatible with all local LLM providers.☆175Updated 9 months ago
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM …☆573Updated 4 months ago
- FastMLX is a high performance production ready API to host MLX models.☆308Updated 3 months ago
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆174Updated last month
- A ggml (C++) re-implementation of tortoise-tts☆187Updated 10 months ago
- Vercel and web-llm template to run wasm models directly in the browser.☆154Updated last year
- API Server for Transformer Lab☆66Updated this week
- A lightweight recreation of OS1/Samantha from the movie Her, running locally in the browser☆101Updated 2 months ago
- Guaranteed Structured Output from any Language Model via Hierarchical State Machines☆139Updated 3 weeks ago
- An innovative library for efficient LLM inference via low-bit quantization☆349Updated 9 months ago
- Run LLMs in the Browser with MLC / WebLLM ✨☆135Updated 8 months ago
- WebAssembly binding for llama.cpp - Enabling on-browser LLM inference☆750Updated 2 weeks ago
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 4 months ago
- Fast Streaming TTS with Orpheus + WebRTC (with FastRTC)☆294Updated 2 months ago
- Train your own small bitnet model☆72Updated 8 months ago
- LLM Inference on consumer devices☆119Updated 3 months ago
- Distributed Inference for mlx LLm☆93Updated 10 months ago
- Locally running LLM with internet access☆95Updated last week
- Realtime tts reading of large textfiles by your favourite voice. +Translation via LLM (Python script)☆52Updated 8 months ago