Picovoice / picollm
On-device LLM Inference Powered by X-Bit Quantization
☆189Updated last week
Related projects ⓘ
Alternatives and complementary repositories for picollm
- Using FastChat-T5 Large Language Model, Vosk API for automatic speech recognition, and Piper for text-to-speech☆111Updated last year
- Recipes for on-device voice AI and local LLM☆66Updated this week
- Realtime tts reading of large textfiles by your favourite voice. +Translation via LLM (Python script)☆47Updated last month
- A fast batching API to serve LLM models☆172Updated 6 months ago
- An AI assistant beyond the chat box.☆315Updated 8 months ago
- On-device streaming text-to-speech engine powered by deep learning☆56Updated 2 weeks ago
- Open source LLM UI, compatible with all local LLM providers.☆167Updated 2 months ago
- Locally running LLM with internet access☆92Updated last month
- Running a LLM on the ESP32☆44Updated last month
- Pybind11 bindings for Whisper.cpp☆45Updated 2 weeks ago
- A multimodal, function calling powered LLM webui.☆208Updated last month
- ☆112Updated this week
- A ggml (C++) re-implementation of tortoise-tts☆159Updated 3 months ago
- ☆191Updated 5 months ago
- A python application that routes incoming prompts to an LLM by category, and can support a single incoming connection from a front end to…☆167Updated this week
- automatically quant GGUF models☆140Updated this week
- A mobile Implementation of llama.cpp☆293Updated 9 months ago
- ☆82Updated this week
- Scripts to create your own moe models using mlx☆86Updated 8 months ago
- idea: https://github.com/nyxkrage/ebook-groupchat/☆82Updated 3 months ago
- Easily view and modify JSON datasets for large language models☆62Updated last month
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆40Updated last month
- Gradio based tool to run opensource LLM models directly from Huggingface☆87Updated 4 months ago
- Fast parallel LLM inference for MLX☆149Updated 4 months ago
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆55Updated last week
- ☆128Updated this week
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAI☆221Updated 6 months ago
- Open Source Application for Advanced LLM Engineering: interact, train, fine-tune, and evaluate large language models on your own computer…☆416Updated this week
- A comprehensive platform for managing, testing, and leveraging Ollama AI models with advanced features for customization, workflow automa…☆45Updated last month
- ezlocalai is an easy to set up local artificial intelligence server with OpenAI Style Endpoints.☆76Updated last week