akx / ggifyLinks
Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp
☆170Updated 9 months ago
Alternatives and similar repositories for ggify
Users that are interested in ggify are comparing it to the libraries listed below
Sorting:
- Download models from the Ollama library, without Ollama☆122Updated last year
- LLM inference in C/C++☆104Updated last week
- automatically quant GGUF models☆219Updated last month
- ☆109Updated 5 months ago
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆266Updated 11 months ago
- LLaVA server (llama.cpp).☆183Updated 2 years ago
- ☆166Updated 6 months ago
- Gemma 2 optimized for your local machine.☆378Updated last year
- LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.☆130Updated 2 years ago
- Practical and advanced guide to LLMOps. It provides a solid understanding of large language models’ general concepts, deployment techniqu…☆78Updated last year
- Falcon LLM ggml framework with CPU and GPU support☆249Updated 2 years ago
- FineTune LLMs in few lines of code (Text2Text, Text2Speech, Speech2Text)☆246Updated 2 years ago
- Maybe the new state of the art vision model? we'll see 🤷♂️☆171Updated 2 years ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated 2 years ago
- A fast batching API to serve LLM models☆189Updated last year
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆119Updated last year
- ☆93Updated last year
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆76Updated 2 years ago
- Distributed Inference for mlx LLm☆100Updated last year
- Low-Rank adapter extraction for fine-tuned transformers models☆180Updated last year
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first app…☆169Updated 2 years ago
- ☆38Updated last year
- ☆68Updated last year
- Access the Cohere Command R family of models☆38Updated 10 months ago
- Extract structured data from local or remote LLM models☆54Updated last year
- ⚡️🧪 Fast LLM Tool Calling Experimentation, big and smol☆158Updated last year
- Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon☆273Updated 2 months ago
- Distribute and run llamafile/LLMs with a single docker image.☆74Updated 8 months ago
- GRDN.AI app for garden optimization☆69Updated 2 months ago