akx / ollama-dlLinks
Download models from the Ollama library, without Ollama
☆108Updated last year
Alternatives and similar repositories for ollama-dl
Users that are interested in ollama-dl are comparing it to the libraries listed below
Sorting:
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆161Updated 6 months ago
- LLM Benchmark for Throughput via Ollama (Local LLMs)☆308Updated 3 months ago
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆119Updated last year
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆214Updated 2 months ago
- Wraps any OpenAI API interface as Responses with MCPs support so it supports Codex. Adding any missing stateful features. Ollama and Vllm…☆126Updated last week
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆98Updated 4 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆82Updated 2 weeks ago
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 9 months ago
- automatically quant GGUF models☆214Updated 2 weeks ago
- ☆106Updated 2 months ago
- Smart proxy for LLM APIs that enables model-specific parameter control, automatic mode switching (like Qwen3's /think and /no_think), and…☆51Updated 5 months ago
- LM inference server implementation based on *.cpp.☆289Updated 2 months ago
- LLM inference in C/C++☆103Updated this week
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆265Updated 8 months ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆63Updated 2 years ago
- ☆207Updated 2 months ago
- No-messing-around sh client for llama.cpp's server☆30Updated last year
- Code execution utilities for Open WebUI & Ollama☆304Updated last year
- LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.☆130Updated 2 years ago
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆84Updated last year
- Distributed Inference for mlx LLm☆98Updated last year
- ☆85Updated last week
- Link you Ollama models to LM-Studio☆146Updated last year
- Prompt Jinja2 templates for LLMs☆34Updated 4 months ago
- A open webui function for better R1 experience☆77Updated 8 months ago
- Gemma 2 optimized for your local machine.☆377Updated last year
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆280Updated 4 months ago
- ☆124Updated last year
- Python package wrapping llama.cpp for on-device LLM inference☆94Updated last month
- A platform to self-host AI on easy mode☆173Updated this week