zhuzilin / faster-nougatLinks
Implementation of nougat that focuses on processing pdf locally.
☆81Updated 5 months ago
Alternatives and similar repositories for faster-nougat
Users that are interested in faster-nougat are comparing it to the libraries listed below
Sorting:
- ☆87Updated 4 months ago
- Minimal, clean code implementation of RAG with mlx using gguf model weights☆51Updated last year
- A framework for evaluating function calls made by LLMs☆37Updated 11 months ago
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆85Updated this week
- Distributed Inference for mlx LLm☆93Updated 10 months ago
- ☆63Updated last month
- Very basic framework for composable parameterized large language model (Q)LoRA / (Q)Dora fine-tuning using mlx, mlx_lm, and OgbujiPT.☆40Updated last week
- ☆66Updated last year
- MLX Swift implementation of Andrej Karpathy's Let's build GPT video☆58Updated last year
- Shared personal notes created while working with the Apple MLX machine learning framework☆24Updated last month
- KAN (Kolmogorov–Arnold Networks) in the MLX framework for Apple Silicon☆18Updated last month
- An example implementation of RLHF (or, more accurately, RLAIF) built on MLX and HuggingFace.☆29Updated last year
- ☆38Updated last year
- 🚀 Scale your RAG pipeline using Ragswift: A scalable centralized embeddings management platform☆38Updated last year
- A simple MLX implementation for pretraining LLMs on Apple Silicon.☆80Updated last month
- For inferring and serving local LLMs using the MLX framework☆104Updated last year
- A collection of tools for your LLMs that run on Modal☆19Updated 3 months ago
- GenAI & agent toolkit for Apple Silicon Mac, implementing JSON schema-steered structured output (3SO) and tool-calling in Python. For mor…☆126Updated 3 weeks ago
- run embeddings in MLX☆90Updated 8 months ago
- A stable, fast and easy-to-use inference library with a focus on a sync-to-async API☆45Updated 8 months ago
- Using modal.com to process FineWeb-edu data☆20Updated 2 months ago
- Scripts to create your own moe models using mlx☆90Updated last year
- ☆114Updated 6 months ago
- Score LLM pretraining data with classifiers☆55Updated last year
- mlx implementations of various transformers, speedups, training☆34Updated last year
- tiny_fnc_engine is a minimal python library that provides a flexible engine for calling functions extracted from a LLM.☆38Updated 9 months ago
- A collection of optimizers for MLX☆36Updated 3 weeks ago
- Data preparation code for CrystalCoder 7B LLM☆45Updated last year
- an implementation of Self-Extend, to expand the context window via grouped attention☆119Updated last year
- Fast parallel LLM inference for MLX☆192Updated 11 months ago