smpanaro / more-ane-transformersLinks
Run transformers (incl. LLMs) on the Apple Neural Engine.
☆62Updated last year
Alternatives and similar repositories for more-ane-transformers
Users that are interested in more-ane-transformers are comparing it to the libraries listed below
Sorting:
- CLI to demonstrate running a large language model (LLM) on Apple Neural Engine.☆114Updated 7 months ago
- FlashAttention (Metal Port)☆512Updated 10 months ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆275Updated last month
- Benchmark of Apple MLX operations on all Apple Silicon chips (GPU, CPU) + MPS and CUDA.☆191Updated 2 months ago
- ☆183Updated 4 months ago
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆187Updated this week
- Tool for exporting Apple Neural Engine-accelerated versions of transformers models on HuggingFace Hub.☆13Updated 2 years ago
- ☆55Updated 2 years ago
- Large Language Models (LLMs) applications and tools running on Apple Silicon in real-time with Apple MLX.☆451Updated 6 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆217Updated last year
- MLX Swift implementation of Andrej Karpathy's Let's build GPT video☆58Updated last year
- Python tools for WhisperKit: Model conversion, optimization and evaluation☆221Updated last week
- See the device (CPU/GPU/ANE) and estimated cost for every layer in your CoreML model.☆22Updated last year
- Explore a simple example of utilizing MLX for RAG application running locally on your Apple Silicon device.☆172Updated last year
- FastMLX is a high performance production ready API to host MLX models.☆320Updated 4 months ago
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆270Updated 10 months ago
- Tool for visual profiling Core ML models, compatible with both package and compiled versions, including reasons for unsupported operation…☆34Updated last year
- mlx image models for Apple Silicon machines☆82Updated 3 months ago
- Fast parallel LLM inference for MLX☆204Updated last year
- run embeddings in MLX☆90Updated 10 months ago
- A multi-platform SwiftUI frontend for running local LLMs with Apple's MLX framework.☆417Updated 9 months ago
- Swift implementation of Flux.1 using mlx-swift☆95Updated 3 weeks ago
- ☆75Updated 8 months ago
- For inferring and serving local LLMs using the MLX framework☆107Updated last year
- Local ML voice chat using high-end models.☆174Updated last month
- ☆23Updated 2 years ago
- MLX Model Manager unifies loading and inferencing with LLMs and VLMs.☆98Updated 6 months ago
- Port of Suno's Bark TTS transformer in Apple's MLX Framework☆83Updated last year
- The easiest way to run the fastest MLX-based LLMs locally☆295Updated 9 months ago
- Start a server from the MLX library.☆189Updated last year