akx / ggifyLinks
Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp
☆154Updated 2 months ago
Alternatives and similar repositories for ggify
Users that are interested in ggify are comparing it to the libraries listed below
Sorting:
- Falcon LLM ggml framework with CPU and GPU support☆246Updated last year
- ☆66Updated last year
- Download models from the Ollama library, without Ollama☆89Updated 8 months ago
- LLaVA server (llama.cpp).☆180Updated last year
- ☆157Updated last year
- A fast batching API to serve LLM models☆183Updated last year
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated last year
- Maybe the new state of the art vision model? we'll see 🤷♂️☆165Updated last year
- 1.58 Bit LLM on Apple Silicon using MLX☆214Updated last year
- An endpoint server for efficiently serving quantized open-source LLMs for code.☆55Updated last year
- Low-Rank adapter extraction for fine-tuned transformers models☆173Updated last year
- Distributed Inference for mlx LLm☆93Updated 11 months ago
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆116Updated last year
- ☆38Updated last year
- GRDN.AI app for garden optimization☆70Updated last year
- Fast parallel LLM inference for MLX☆198Updated last year
- automatically quant GGUF models☆187Updated this week
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆75Updated last year
- LLM inference in C/C++☆78Updated 3 weeks ago
- Minimal, clean code implementation of RAG with mlx using gguf model weights☆52Updated last year
- For inferring and serving local LLMs using the MLX framework☆105Updated last year
- LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.☆127Updated 2 years ago
- Gemma 2 optimized for your local machine.☆376Updated 11 months ago
- Inference of Large Multimodal Models in C/C++. LLaVA and others☆47Updated last year
- Easily convert HuggingFace models to GGUF-format for llama.cpp☆21Updated 11 months ago
- Python bindings for ggml☆142Updated 10 months ago
- Scripts to create your own moe models using mlx☆90Updated last year
- Unsloth Studio☆93Updated 3 months ago
- LLM-powered lossless compression tool☆283Updated 11 months ago
- Train your own small bitnet model☆74Updated 8 months ago