ggml-org / ciLinks
CI for ggml and related projects
☆31Updated 2 months ago
Alternatives and similar repositories for ci
Users that are interested in ci are comparing it to the libraries listed below
Sorting:
- LLM inference in C/C++☆102Updated 2 weeks ago
- Command line tool for Deep Infra cloud ML inference service☆33Updated last year
- Granite 3.1 Language Models☆131Updated 4 months ago
- A minimalistic C++ Jinja templating engine for LLM chat templates☆197Updated 2 months ago
- 1.58 Bit LLM on Apple Silicon using MLX☆225Updated last year
- Utility library to work with character cards and roleplay AI in general☆44Updated 2 years ago
- Transformer GPU VRAM estimator☆66Updated last year
- ☆163Updated 3 months ago
- ☆18Updated 11 months ago
- C API for MLX☆151Updated this week
- Deploy your GGML models to HuggingFace Spaces with Docker and gradio☆38Updated 2 years ago
- Examples of integrating the OpenRouter API☆245Updated this week
- ☆67Updated last year
- Gemma 2 optimized for your local machine.☆377Updated last year
- MLX support for the Open Neural Network Exchange (ONNX)☆62Updated last year
- 🔊 We believe in a future where developers are amplified, not automated☆114Updated last month
- For individual users, watsonx Code Assistant can access a local IBM Granite model☆37Updated 4 months ago
- LLM-based code completion engine☆190Updated 10 months ago
- ☆166Updated last year
- Thin wrapper around GGML to make life easier☆40Updated 2 weeks ago
- Public reports detailing responses to sets of prompts by Large Language Models.☆32Updated 10 months ago
- ☆116Updated 11 months ago
- Resources regarding evML (edge verified machine learning)☆19Updated 10 months ago
- An endpoint server for efficiently serving quantized open-source LLMs for code.☆58Updated 2 years ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆162Updated 6 months ago
- AirLLM 70B inference with single 4GB GPU☆14Updated 4 months ago
- ☆102Updated last year
- Fast parallel LLM inference for MLX☆232Updated last year
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆227Updated 3 weeks ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆53Updated last year