akx / ollama-dl
Download models from the Ollama library, without Ollama
☆66Updated 4 months ago
Alternatives and similar repositories for ollama-dl:
Users that are interested in ollama-dl are comparing it to the libraries listed below
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆73Updated 3 months ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆127Updated 6 months ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆65Updated last year
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.☆115Updated 10 months ago
- SiLLM simplifies the process of training and running Large Language Models (LLMs) on Apple Silicon by leveraging the MLX framework.☆260Updated 2 weeks ago
- klmbr - a prompt pre-processing technique to break through the barrier of entropy while generating text with LLMs☆71Updated 6 months ago
- A guidance compatibility layer for llama-cpp-python☆34Updated last year
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆55Updated last month
- Extract structured data from local or remote LLM models☆41Updated 9 months ago
- ☆197Updated 2 weeks ago
- ☆84Updated 9 months ago
- Yet Another (LLM) Web UI, made with Gemini☆11Updated 3 months ago
- Gradio based tool to run opensource LLM models directly from Huggingface☆91Updated 9 months ago
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆181Updated 2 months ago
- MLX-Embeddings is the best package for running Vision and Language Embedding models locally on your Mac using MLX.☆112Updated this week
- Forces DeepSeek R1 models to engage in extended reasoning by intercepting early termination tokens.☆19Updated last month
- ☆31Updated last year
- Tcurtsni: Reverse Instruction Chat, ever wonder what your LLM wants to ask you?☆21Updated 9 months ago
- An endpoint server for efficiently serving quantized open-source LLMs for code.☆54Updated last year
- Serving LLMs in the HF-Transformers format via a PyFlask API☆71Updated 6 months ago
- Local LLM inference & management server with built-in OpenAI API☆31Updated 11 months ago
- GRDN.AI app for garden optimization☆70Updated last year
- LLM plugin providing access to models running on an Ollama server☆261Updated 3 weeks ago
- Lightweight Inference server for OpenVINO☆143Updated this week
- RetroChat is a powerful command-line interface for interacting with various AI language models. It provides a seamless experience for eng…☆72Updated 3 months ago
- Generate Your Own Private Morning Radio for Commute☆34Updated last month
- ☆45Updated 10 months ago
- Gemma 2 optimized for your local machine.☆366Updated 7 months ago
- Distributed Inference for mlx LLm☆87Updated 7 months ago
- automatically quant GGUF models☆164Updated last week