smpanaro / more-ane-transformersLinks
Run transformers (incl. LLMs) on the Apple Neural Engine.
☆63Updated 2 years ago
Alternatives and similar repositories for more-ane-transformers
Users that are interested in more-ane-transformers are comparing it to the libraries listed below
Sorting:
- CLI to demonstrate running a large language model (LLM) on Apple Neural Engine.☆117Updated last year
- ☆58Updated 2 years ago
- Tool for exporting Apple Neural Engine-accelerated versions of transformers models on HuggingFace Hub.☆13Updated 2 years ago
- FlashAttention (Metal Port)☆572Updated last year
- See the device (CPU/GPU/ANE) and estimated cost for every layer in your CoreML model.☆24Updated 2 months ago
- ☆24Updated 2 years ago
- Tool for visual profiling Core ML models, compatible with both package and compiled versions, including reasons for unsupported operation…☆35Updated last year
- Benchmark of Apple MLX operations on all Apple Silicon chips (GPU, CPU) + MPS and CUDA.☆209Updated 7 months ago
- ☆196Updated 9 months ago
- Local ML voice chat using high-end models.☆178Updated 3 weeks ago
- run embeddings in MLX☆96Updated last year
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆244Updated 2 months ago
- Python tools for WhisperKit: Model conversion, optimization and evaluation☆234Updated 2 months ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆286Updated 6 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆234Updated last year
- MLX Swift implementation of Andrej Karpathy's Let's build GPT video☆62Updated last year
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.☆459Updated 11 months ago
- Explore a simple example of utilizing MLX for RAG application running locally on your Apple Silicon device.☆179Updated last year
- Port of Suno's Bark TTS transformer in Apple's MLX Framework☆86Updated last year
- FastMLX is a high performance production ready API to host MLX models.☆339Updated 9 months ago
- For inferring and serving local LLMs using the MLX framework☆109Updated last year
- ☆126Updated 6 months ago
- mlx image models for Apple Silicon machines☆90Updated last month
- Fast parallel LLM inference for MLX☆241Updated last year
- A multi-platform SwiftUI frontend for running local LLMs with Apple's MLX framework.☆429Updated last year
- ☆76Updated last year
- Swift implementation of Flux.1 using mlx-swift☆113Updated 5 months ago
- An LLM agnostic desktop and mobile client.☆308Updated 3 months ago
- ☆306Updated 8 months ago
- MLX Model Manager unifies loading and inferencing with LLMs and VLMs.☆104Updated 11 months ago