senstella / parakeet-mlx
An implementation of the Nvidia's Parakeet models for Apple Silicon using MLX.
☆84Updated this week
Alternatives and similar repositories for parakeet-mlx:
Users that are interested in parakeet-mlx are comparing it to the libraries listed below
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆147Updated 3 weeks ago
- MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. I…☆359Updated last week
- An implementation of the CSM(Conversation Speech Model) for Apple Silicon using MLX.☆331Updated last week
- FastMLX is a high performance production ready API to host MLX models.☆297Updated last month
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆78Updated 5 months ago
- Blazing fast whisper turbo for ASR (speech-to-text) tasks☆205Updated 6 months ago
- Python tools for WhisperKit: Model conversion, optimization and evaluation☆212Updated this week
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆264Updated this week
- For inferring and serving local LLMs using the MLX framework☆103Updated last year
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆265Updated 8 months ago
- Guaranteed Structured Output from any Language Model via Hierarchical State Machines☆128Updated last week
- ☆91Updated last month
- run embeddings in MLX☆88Updated 7 months ago
- ☆168Updated last month
- Grammar checker with a keyboard shortcut for Ollama and Apple MLX with Automator on macOS.☆80Updated last year
- GenAI & agent toolkit for Apple Silicon Mac, implementing JSON schema-steered structured output (3SO) and tool-calling in Python. For mor…☆123Updated 2 months ago
- Distributed Inference for mlx LLm☆91Updated 9 months ago
- Fast parallel LLM inference for MLX☆187Updated 10 months ago
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆115Updated 11 months ago
- A Conversational Speech Generation Model with Gradio UI and OpenAI compatible API. UI and API support CUDA, MLX and CPU devices.☆181Updated this week
- Fast Streaming TTS with Orpheus + WebRTC (with FastRTC)☆274Updated last month
- Scripts to create your own moe models using mlx☆89Updated last year
- Port of Suno's Bark TTS transformer in Apple's MLX Framework☆81Updated last year
- mlx implementations of various transformers, speedups, training☆34Updated last year
- Optimized Ollama LLM server configuration for Mac Studio and other Apple Silicon Macs. Headless setup with automatic startup, resource op…☆176Updated 2 months ago
- MLX Model Manager unifies loading and inferencing with LLMs and VLMs.☆89Updated 3 months ago
- Run Orpheus 3B Locally With LM Studio☆396Updated last month
- Start a server from the MLX library.☆185Updated 9 months ago
- Embarrassingly Easy Fully Non-Autoregressive Zero-Shot TTS (E2 TTS) in MLX☆27Updated 6 months ago
- MLX-based QA pair generator and LLM finetuning tool in Streamlit☆33Updated 5 months ago